File match vector.json vector.toml vector.yaml
Schema URL https://catalog.lintel.tools/schemas/schemastore/vector/latest.json
Source https://www.schemastore.org/vector.json

Validate with Lintel

npx @lintel/lintel check

A complete Vector configuration.

All of

1. object object
api object

API options.

Default:
{
  "enabled": false,
  "address": "127.0.0.1:8686",
  "playground": true
}
4 nested properties
address null | string

The socket address to listen on for the API endpoint.

Default: "127.0.0.1:8686"
enabled boolean

Whether or not the API endpoint is available.

Default: false
graphql boolean

Whether or not the GraphQL endpoint is enabled

Default: true
playground boolean

Whether or not to expose the GraphQL playground on the API endpoint.

Default: true
enrichment_tables Record<string, any>

All configured enrichment tables.

Default:
{}
healthchecks object

Healthcheck options.

Default:
{
  "enabled": true,
  "require_healthy": false
}
2 nested properties
enabled boolean

Can be overridden on a per-sink basis.

require_healthy boolean

When enabled and a sink reports not being healthy, Vector will exit during start-up.

Can be alternatively set, and overridden by, the --require-healthy command-line flag.

provider null

Configuration providers allow sourcing configuration information from a source other than the typical configuration files that must be passed to Vector.

schema object

Schema options.

Default:
{
  "enabled": false,
  "validation": false,
  "log_namespace": null
}
3 nested properties
enabled boolean

Whether or not schema is enabled.

Default: false
log_namespace boolean | null

Whether or not to enable log namespacing.

validation boolean

Whether or not schema validation is enabled.

Default: false
secret Record<string, any>

All configured secrets backends.

Default:
{}

All configured sinks.

Default:
{}
sources Record<string, vector::config::source::SourceOuter>

All configured sources.

Default:
{}
tests object[]

All configured unit tests.

Default:
[]

All configured transforms.

Default:
{}
2. object object

Global configuration options.

acknowledgements object

See End-to-end Acknowledgements for more information on how event acknowledgement is handled.

1 nested properties
enabled boolean | null

When enabled for a sink, any source connected to that sink, where the source supports end-to-end acknowledgements as well, waits for events to be acknowledged by all connected sinks before acknowledging them at the source.

Enabling or disabling acknowledgements at the sink level takes precedence over any global acknowledgements configuration.

This is the directory where Vector will store any state data, such as disk buffers, file checkpoints, and more.

Vector must have write permissions to this directory.

Default: "/var/lib/vector/"
expire_metrics null | object

Deprecated: use expire_metrics_secs instead

Default: null
expire_metrics_secs number | null

Set this to a value larger than your internal_metrics scrape interval (default 5 minutes) that metrics live long enough to be emitted and captured,

min=-9007199254740990.0max=9007199254740990.0
log_schema object

This is used if a component does not have its own specific log schema. All events use a log schema, whether or not the default is used, to assign event fields on incoming events.

Default:
{
  "message_key": ".message",
  "timestamp_key": ".timestamp",
  "host_key": ".host",
  "source_type_key": ".source_type",
  "metadata_key": ".metadata"
}
5 nested properties
host_key string

An optional path that deserializes an empty string to None.

message_key string

An optional path that deserializes an empty string to None.

metadata_key string

An optional path that deserializes an empty string to None.

source_type_key string

An optional path that deserializes an empty string to None.

timestamp_key string

An optional path that deserializes an empty string to None.

proxy object

Configure to proxy traffic through an HTTP(S) proxy when making external requests.

Similar to common proxy configuration convention, you can set different proxies to use based on the type of traffic being proxied. You can also set specific hosts that should not be proxied.

4 nested properties
enabled boolean

Enables proxying support.

Default: true
http string | null

Must be a valid URI string.

Default: null
format=uri
https string | null

Must be a valid URI string.

Default: null
format=uri
no_proxy string[]

Multiple patterns are allowed:

PatternExample match
Domain namesexample.com matches requests to example.com
Wildcard domains.example.com matches requests to example.com and its subdomains
IP addresses127.0.0.1 matches requests to 127.0.0.1
CIDR blocks192.168.0.0/16 matches requests to any IP addresses in this range
Splat* matches all hosts
Default:
[]
telemetry object

Determines whether source and service tags should be emitted with the component_sent_* and component_received_* events.

Default:
{
  "tags": {
    "emit_service": false,
    "emit_source": false
  }
}
1 nested properties
tags object

Configures whether to emit certain tags

Default:
{
  "emit_service": false,
  "emit_source": false
}
2 nested properties
emit_service boolean

True if the service tag should be emitted in the component_received_* and component_sent_* telemetry.

emit_source boolean

True if the source tag should be emitted in the component_received_* and component_sent_* telemetry.

The time zone name may be any name in the TZ database or local to indicate system local time.

Note that in Vector/VRL all timestamps are represented in UTC.

Default: null

Definitions

codecs::MetricTagValues const: "single" | const: "full"

The user configuration to choose the metric tag strategy.

codecs::common::length_delimited::LengthDelimitedCoderOptions object

Options for building a LengthDelimitedDecoder or LengthDelimitedEncoder.

length_field_is_big_endian boolean

Length field byte order (little or big endian)

Default: true
length_field_length integer

Number of bytes representing the field length

Default: 4
min=0.0max=9007199254740990.0
length_field_offset integer

Number of bytes in the header before the length field

Default: 0
min=0.0max=9007199254740990.0
max_frame_length integer

Maximum frame length

Default: 8388608
min=0.0max=9007199254740990.0
codecs::decoding::DeserializerConfig object | object | object

Configures how events are decoded from raw bytes.

codecs::decoding::FramingConfig object

Framing handles how events are separated when encoded in a raw byte form, where each event is a frame that must be prefixed, or delimited, in a way that marks where an event begins and ends within the byte stream.

codecs::encoding::format::csv::CsvSerializerConfig object

Config used to build a CsvSerializer.

csv object required

The CSV Serializer Options.

7 nested properties

If a field is not present in the event, the output will be an empty string.

Values of type Array, Object, and Regex are not supported and the output will be an empty string.

capacity integer

Set the capacity (in bytes) of the internal buffer used in the CSV writer. This defaults to a reasonable setting.

Default: 8192
min=0.0max=9007199254740990.0
delimiter integer

The field delimiter to use when writing CSV.

Default: 44
min=0.0max=255.0
double_quote boolean

This is enabled by default, but it may be disabled. When disabled, quotes in field data are escaped instead of doubled.

Default: true
escape integer

In some variants of CSV, quotes are escaped using a special escape character like \ (instead of escaping quotes by doubling them).

To use this, double_quotes needs to be disabled as well otherwise it is ignored.

Default: 34
min=0.0max=255.0
quote integer

The quote character to use when writing CSV.

Default: 34
min=0.0max=255.0
quote_style const: "always" | const: "necessary" | const: "non_numeric" | const: "never"

The quoting style to use when writing CSV data.

Default: "necessary"
codecs::encoding::format::json::JsonSerializerConfig object

Config used to build a JsonSerializer.

json object

Options for the JsonSerializer.

Default:
{
  "pretty": false
}
1 nested properties
pretty boolean

Whether to use pretty JSON formatting.

Default: false
metric_tag_values const: "single" | const: "full"

The user configuration to choose the metric tag strategy.

core::option::Option<codecs::decoding::FramingConfig> null | object

Framing handles how events are separated when encoded in a raw byte form, where each event is a frame that must be prefixed, or delimited, in a way that marks where an event begins and ends within the byte stream.

core::option::Option<std::path::PathBuf> null | stdlib::PathBuf

A file path.

core::option::Option<vector::aws::region::RegionOrEndpoint> null | vector::aws::region::RegionOrEndpoint

Configuration of the region/endpoint to use when interacting with an AWS service.

core::option::Option<vector::conditions::AnyCondition> null | string | object | object | object

Many methods exist for matching events, such as using a VRL expression, a Datadog Search query string, or hard-coded matchers like "must be a metric" or "fields A, B, and C must match these constraints".

As VRL is the most common way to apply conditions to events, this type provides a shortcut to define VRL expressions directly in the configuration by passing the VRL expression as a string:

condition = '.message == "hooray"'

When other condition types are required, they can be specified with an enum-style notation:

condition.type = 'datadog_search'
condition.source = 'NOT "foo"'
core::option::Option<vector::http::Auth> null | object | object

HTTP authentication should be used with HTTPS only, as the authentication credentials are passed as an HTTP header without any additional encryption beyond what is provided by the transport itself.

core::option::Option<vector::nats::NatsAuthConfig> null | object | object | object | object

Configuration of the authentication strategy when interacting with NATS.

core::option::Option<vector::sinks::influxdb::InfluxDb1Settings> null | object

Configuration settings for InfluxDB v0.x/v1.x.

core::option::Option<vector::sinks::influxdb::InfluxDb2Settings> null | object

Configuration settings for InfluxDB v2.x.

core::option::Option<vector::sinks::util::buffer::compression::Compression> null | const: "none" | const: "gzip" | const: "zlib" | const: "zstd" | const: "snappy" | object

All compression algorithms use the default compression level unless otherwise specified.

core::option::Option<vector::sinks::util::uri::UriSerde> null | vector::sinks::util::uri::UriSerde
core::option::Option<vector::sources::util::http::auth::HttpSourceAuthConfig> null | object

HTTP Basic authentication configuration.

core::option::Option<vector::sources::util::multiline_config::MultilineConfig> null | object

Configuration of multi-line aggregation.

core::option::Option<vector::template::Template> null | vector::template::Template

In many cases, components can be configured so that part of the component's functionality can be customized on a per-event basis. For example, you have a sink that writes events to a file and you want to specify which file an event should go to by using an event field as part of the input to the filename used.

By using Template, users can specify either fixed strings or templated strings. Templated strings use a common syntax to refer to fields in an event that is used as the input data when rendering the template. An example of a fixed string is my-file.log. An example of a template string is my-file-{{key}}.log, where {{key}} is the key's value when the template is rendered into a string.

core::option::Option<vector_common::sensitive_string::SensitiveString> null | vector_common::sensitive_string::SensitiveString

Wrapper for sensitive strings containing credentials

core::option::Option<vector_core::ipallowlist::IpAllowlistConfig> null | string[]

List of allowed origin IP networks. IP addresses must be in CIDR notation.

core::option::Option<vector_core::tcp::TcpKeepaliveConfig> null | object

TCP keepalive settings for socket-based components.

core::option::Option<vector_core::tls::settings::TlsConfig> null | vector_core::tls::settings::TlsConfig

TLS configuration.

core::option::Option<vector_core::tls::settings::TlsEnableableConfig> null | vector_core::tls::settings::TlsEnableableConfig

Configures the TLS options for incoming/outgoing connections.

core::option::Option<vector_core::tls::settings::TlsSourceConfig> null

TlsEnableableConfig for sources, adding metadata from the client certificate.

core::option::Option<vector_lookup::lookup_v2::ConfigValuePath> null | vector_lookup::lookup_v2::ConfigValuePath

A wrapper around OwnedValuePath that allows it to be used in Vector config. This requires a valid path to be used. If you want to allow optional paths, use [optional_path::OptionalValuePath].

core::option::Option<vector_lookup::lookup_v2::optional_path::OptionalTargetPath> null | vector_lookup::lookup_v2::optional_path::OptionalTargetPath

An optional path that deserializes an empty string to None.

core::option::Option<vector_lookup::lookup_v2::optional_path::OptionalValuePath> null | vector_lookup::lookup_v2::optional_path::OptionalValuePath

An optional path that deserializes an empty string to None.

core::option::Option<vrl::compiler::datetime::TimeZone> null | const: "local" | string

This can refer to any valid timezone as defined in the TZ database, or "local" which refers to the system local timezone. It will default to the globally configured timezone.

file_source::ReadFromConfig const: "beginning" | const: "end"

File position to use when reading a new file.

serde_with::DurationFractionalSeconds number

A span of time, in fractional seconds.

serde_with::DurationMilliSeconds integer

A span of time, in whole milliseconds.

serde_with::DurationSeconds integer

A span of time, in whole seconds.

stdlib::PathBuf string

A file path.

stdlib::SocketAddr string

An internet socket address, either IPv4 or IPv6.

vector::amqp::AmqpConfig object

AMQP connection options.

connection_string string required

The URI has the format of amqp://<user>:<password>@<host>:<port>/<vhost>?timeout=<seconds>.

The default vhost can be specified by using a value of %2f.

To connect over TLS, a scheme of amqps can be specified instead. For example, amqps://.... Additional TLS settings, such as client certificate verification, can be configured under the tls section.

vector::aws::auth::AwsAuthentication object | object | object | object

Configuration of the authentication strategy for interacting with AWS services.

vector::aws::region::RegionOrEndpoint object

Configuration of the region/endpoint to use when interacting with an AWS service.

endpoint string | null

Custom endpoint for use with AWS-compatible services.

region string | null
vector::codecs::encoding::config::EncodingConfig

Configures how events are encoded into raw bytes.

vector::codecs::encoding::config::EncodingConfigWithFraming object

Encoding configuration.

encoding required

Configures how events are encoded into raw bytes.

All of: object | object | object | object | object | object object | object | object | object | object | object, object object
framing null | object | object

Framing configuration.

vector::codecs::encoding::transformer::Transformer object

Transformations to prepare an event for serialization.

except_fields array | null

List of fields that are excluded from the encoded event.

Default: null
only_fields array | null

List of fields that are included in the encoded event.

Default: null
timestamp_format null | const: "unix" | const: "rfc3339" | const: "unix_ms" | const: "unix_us" | const: "unix_ns" | const: "unix_float"

Format used for timestamp fields.

Default: null
vector::conditions::AnyCondition string | object | object | object

Many methods exist for matching events, such as using a VRL expression, a Datadog Search query string, or hard-coded matchers like "must be a metric" or "fields A, B, and C must match these constraints".

As VRL is the most common way to apply conditions to events, this type provides a shortcut to define VRL expressions directly in the configuration by passing the VRL expression as a string:

condition = '.message == "hooray"'

When other condition types are required, they can be specified with an enum-style notation:

condition.type = 'datadog_search'
condition.source = 'NOT "foo"'
vector::config::TestInput object

An input describes not only the type of event to insert, but also which transform within the configuration to insert it to.

insert_at string required

The name of the transform to insert the input event to.

log_fields Record<string, string | integer | number | boolean>

Only relevant when type is log.

metric null

Only relevant when type is metric.

source string | null

Only relevant when type is vrl.

type string

Can be either raw, vrl, log, or `metric.

Default: "raw"
value string | null

Use this only when the input event should be a raw event (i.e. unprocessed/undecoded log event) and when the input type is set to raw.

vector::config::dot_graph::GraphConfig object

Configure output for component when generated with graph command

node_attributes Record<string, string>

They are added to the node as provided

vector::config::sink::SinkOuter<alloc::string::String>

Fully resolved sink component.

vector::config::source::SourceOuter

Fully resolved source component.

vector::config::transform::TransformOuter<alloc::string::String>

Fully resolved transform component.

vector::config::unit_test::unit_test_components::UnitTestSinkConfig object

Configuration for the unit_test sink.

test_name string required

Name of the test that this sink is being used for.

transform_ids string[] required

List of names of the transform/branch associated with this sink.

vector::config::unit_test::unit_test_components::UnitTestSourceConfig object

Configuration for the unit_test source.

vector::config::unit_test::unit_test_components::UnitTestStreamSinkConfig object

Configuration for the unit_test_stream sink.

vector::config::unit_test::unit_test_components::UnitTestStreamSourceConfig object

Configuration for the unit_test_stream source.

vector::enrichment_tables::file::FileConfig object

Configuration for the file enrichment table.

file object required

File-specific settings.

2 nested properties
encoding object required

File encoding configuration.

path string required

A file path.

pattern=(\/.*|[a-zA-Z]:\\(?:([^<>:"\/\\|?*]*[^<>:"\/\\|?*.]\\|..\\)*([^<>:"\/\\|?*]*[^<>:"\/\\|?*.]\\?|..\\))?)
schema Record<string, string>

This is used to coerce log fields from strings into their proper types. The available types are listed in the Types list below.

Timestamp coercions need to be prefaced with timestamp|, for example "timestamp|%F". Timestamp specifiers can use either of the following:

  1. One of the built-in-formats listed in the Timestamp Formats table below.
  2. The time format specifiers from Rust’s chrono library.

Types

  • bool
  • string
  • float
  • integer
  • date
  • timestamp (see the table below for formats)

Timestamp Formats

FormatDescriptionExample
%F %TYYYY-MM-DD HH:MM:SS2020-12-01 02:37:54
%v %TDD-Mmm-YYYY HH:MM:SS01-Dec-2020 02:37:54
%FT%TISO 8601/RFC 3339, without time zone2020-12-01T02:37:54
%FT%TZISO 8601/RFC 3339, UTC2020-12-01T09:37:54Z
%+ISO 8601/RFC 3339, UTC, with time zone2020-12-01T02:37:54-07:00
%a, %d %b %Y %TRFC 822/RFC 2822, without time zoneTue, 01 Dec 2020 02:37:54
%a %b %e %T %Yctime formatTue Dec 1 02:37:54 2020
%sUNIX timestamp1606790274
%a %d %b %T %Ydate command, without time zoneTue 01 Dec 02:37:54 2020
%a %d %b %T %Z %Ydate command, with time zoneTue 01 Dec 02:37:54 PST 2020
%a %d %b %T %z %Ydate command, with numeric time zoneTue 01 Dec 02:37:54 -0700 2020
%a %d %b %T %#z %Ydate command, with numeric time zone (minutes can be missing or present)Tue 01 Dec 02:37:54 -07 2020
Default:
{}
vector::enrichment_tables::geoip::GeoipConfig object

Configuration for the geoip enrichment table.

path string required

Other databases, such as the country database, are not supported. mmdb enrichment table can be used for other databases.

locale string

MaxMind includes localized versions of some of the fields within their database, such as country name. This setting can control which of those localized versions are returned by the transform.

More information on which portions of the geolocation data are localized, and what languages are available, can be found here.

Default: "en"
vector::enrichment_tables::mmdb::MmdbConfig object

Configuration for the mmdb enrichment table.

path string required
vector::gcp::GcpAuthConfig object

Configuration of the authentication strategy for interacting with GCP services.

Either an API key or a path to a service account credentials JSON file can be specified.

If both are unset, the GOOGLE_APPLICATION_CREDENTIALS environment variable is checked for a filename. If no filename is named, an attempt is made to fetch an instance service account for the compute instance the program is running on. If this is not on a GCE instance, then you must define it with an API key or service account credentials JSON file.

credentials_path string | null

Either an API key or a path to a service account credentials JSON file can be specified.

If both are unset, the GOOGLE_APPLICATION_CREDENTIALS environment variable is checked for a filename. If no filename is named, an attempt is made to fetch an instance service account for the compute instance the program is running on. If this is not on a GCE instance, then you must define it with an API key or service account credentials JSON file.

skip_authentication boolean

Skip all authentication handling. For use with integration tests only.

Default: false
vector::http::KeepaliveConfig object

Configuration of HTTP server keepalive parameters.

max_connection_age_jitter_factor number

A value of 0.1 means that the actual duration will be between 90% and 110% of the specified maximum duration.

Default: 0.1
min=0.0max=1.0
max_connection_age_secs integer | null

Only applies to HTTP/0.9, HTTP/1.0, and HTTP/1.1 requests.

A random jitter configured by max_connection_age_jitter_factor is added to the specified duration to spread out connection storms.

Default: 300
min=0.0max=9007199254740990.0
vector::internal_events::file::FileInternalMetricsConfig object

Configuration of internal metrics for file-based components.

include_file_tag boolean

This is useful for distinguishing between different files while monitoring. However, the tag's cardinality is unbounded.

Default: false
vector::kafka::KafkaAuthConfig object

Kafka authentication configuration.

sasl null | object

Configuration for SASL authentication when interacting with Kafka.

vector::providers::http::HttpConfig

Configuration for the http provider.

vector::secrets::aws_secrets_manager::AwsSecretsManagerBackend

Configuration for the aws_secrets_manager secrets backend.

vector::secrets::directory::DirectoryBackend object

Configuration for the directory secrets backend.

path string required

A file path.

pattern=(\/.*|[a-zA-Z]:\\(?:([^<>:"\/\\|?*]*[^<>:"\/\\|?*.]\\|..\\)*([^<>:"\/\\|?*]*[^<>:"\/\\|?*.]\\?|..\\))?)
remove_trailing_whitespace boolean

Remove trailing whitespace from file contents.

Default: false
vector::secrets::exec::ExecBackend object

Configuration for the exec secrets backend.

command string[] required

The path to the script or binary must be the first argument.

timeout integer

The timeout, in seconds, to wait for the command to complete.

Default: 5
min=0.0max=9007199254740990.0
vector::secrets::file::FileBackend object

Configuration for the file secrets backend.

path string required

A file path.

pattern=(\/.*|[a-zA-Z]:\\(?:([^<>:"\/\\|?*]*[^<>:"\/\\|?*.]\\|..\\)*([^<>:"\/\\|?*]*[^<>:"\/\\|?*.]\\?|..\\))?)
vector::secrets::test::TestBackend object

Configuration for the test secrets backend.

replacement string required

Fixed value to replace all secrets with.

vector::sinks::Sinks

Configurable sinks in Vector.

vector::sinks::amqp::config::AmqpSinkConfig

Supports AMQP version 0.9.1

vector::sinks::appsignal::config::AppsignalConfig object

Configuration for the appsignal sink.

push_api_key string required

Wrapper for sensitive strings containing credentials

acknowledgements object

See End-to-end Acknowledgements for more information on how event acknowledgement is handled.

1 nested properties
enabled boolean | null

When enabled for a sink, any source connected to that sink, where the source supports end-to-end acknowledgements as well, waits for events to be acknowledged by all connected sinks before acknowledging them at the source.

Enabling or disabling acknowledgements at the sink level takes precedence over any global acknowledgements configuration.

batch object

Event batching behavior.

Default:
{
  "max_bytes": null,
  "max_events": null,
  "timeout_secs": null
}
3 nested properties
max_bytes integer | null

This is based on the uncompressed size of the batched events, before they are serialized/compressed.

Default: 450000
min=0.0max=9007199254740990.0
max_events integer | null

The maximum size of a batch before it is flushed.

Default: 100
min=0.0max=9007199254740990.0
timeout_secs number | null

The maximum age of a batch before it is flushed.

Default: 1.0
min=-9007199254740990.0max=9007199254740990.0
compression const: "none" | const: "gzip" | const: "zlib" | const: "zstd" | const: "snappy" | object

All compression algorithms use the default compression level unless otherwise specified.

encoding object

Transformations to prepare an event for serialization.

3 nested properties
except_fields array | null

List of fields that are excluded from the encoded event.

Default: null
only_fields array | null

List of fields that are included in the encoded event.

Default: null
timestamp_format null | const: "unix" | const: "rfc3339" | const: "unix_ms" | const: "unix_us" | const: "unix_ns" | const: "unix_float"

Format used for timestamp fields.

Default: null
endpoint string

The URI for the AppSignal API to send data to.

Default: "https://appsignal-endpoint.net"
format=uri
request object

Various settings can be configured, such as concurrency and rate limits, timeouts, retry behavior, etc.

Note that the retry backoff policy follows the Fibonacci sequence.

9 nested properties
adaptive_concurrency object

These parameters typically do not require changes from the default, and incorrect values can lead to meta-stable or unstable performance and sink behavior. Proceed with caution.

5 nested properties
decrease_ratio number

Valid values are greater than 0 and less than 1. Smaller values cause the algorithm to scale back rapidly when latency increases.

Note that the new limit is rounded down after applying this ratio.

Default: 0.9
min=0.0max=1.0
ewma_alpha number

Valid values are greater than 0 and less than 1.

ARC uses an exponentially weighted moving average (EWMA) of past RTT measurements as a reference to compare with the current RTT. Smaller values cause this reference to adjust more slowly, which may be useful if a service has unusually high response variability.

Default: 0.4
min=0.0max=1.0
initial_concurrency integer

It is recommended to set this value to your service's average limit if you're seeing that it takes a long time to ramp up adaptive concurrency after a restart. You can find this value by looking at the adaptive_concurrency_limit metric.

Default: 1
min=1.0max=9007199254740990.0
max_concurrency_limit integer

The adaptive request concurrency limit will not go above this bound. This is put in place as a safeguard.

Default: 200
min=1.0max=9007199254740990.0
rtt_deviation_scale number

Valid values are greater than or equal to 0, and we expect reasonable values to range from 1.0 to 3.0.

When calculating the past RTT average, we also compute a secondary “deviation” value that indicates how variable those values are. We use that deviation when comparing the past RTT average to the current measurements, so we can ignore increases in RTT that are within an expected range. This factor is used to scale up the deviation to an appropriate range. Larger values cause the algorithm to ignore larger increases in the RTT.

Default: 2.5
min=0.0max=9007199254740990.0
concurrency const: "none" | const: "adaptive" | integer

Configuration for outbound request concurrency.

This can be set either to one of the below enum values or to a positive integer, which denotes a fixed concurrency limit.

rate_limit_duration_secs integer

The time window used for the rate_limit_num option.

Default: 1
min=0.0max=9007199254740990.0
rate_limit_num integer

The maximum number of requests allowed within the rate_limit_duration_secs time window.

Default: 9223372036854775807
min=0.0max=9007199254740990.0
retry_attempts integer

The maximum number of retries to make for failed requests.

Default: 9223372036854775807
min=0.0max=9007199254740990.0
retry_initial_backoff_secs integer

After the first retry has failed, the fibonacci sequence is used to select future backoffs.

Default: 1
min=1.0max=9007199254740990.0
retry_jitter_mode const: "None" | const: "Full"

The jitter mode to use for retry backoff behavior.

retry_max_duration_secs integer

The maximum amount of time to wait between retries.

Default: 30
min=1.0max=9007199254740990.0
timeout_secs integer

Datadog highly recommends that you do not lower this value below the service's internal timeout, as this could create orphaned requests, pile on retries, and result in duplicate data downstream.

Default: 60
min=0.0max=9007199254740990.0
vector::sinks::aws_cloudwatch_logs::config::CloudwatchLogsSinkConfig

Configuration for the aws_cloudwatch_logs sink.

vector::sinks::aws_cloudwatch_metrics::CloudWatchMetricsSinkConfig

Configuration for the aws_cloudwatch_metrics sink.

vector::sinks::aws_kinesis::config::KinesisSinkBaseConfig

Base configuration for the aws_kinesis_ sinks. The actual specific sink configuration types should either wrap this in a newtype wrapper, or should extend it in a new struct with serde(flatten).

vector::sinks::aws_kinesis::firehose::config::KinesisFirehoseSinkConfig

Configuration for the aws_kinesis_firehose sink.

vector::sinks::aws_kinesis::streams::config::KinesisStreamsSinkConfig

Configuration for the aws_kinesis_streams sink.

vector::sinks::aws_s3::config::S3SinkConfig

Configuration for the aws_s3 sink.

vector::sinks::aws_s_s::config::BaseSSSinkConfig object

Base Configuration aws_s_s for sns and sqs sink.

encoding required

Configures how events are encoded into raw bytes.

All of: object | object | object | object | object | object object | object | object | object | object | object, object object
acknowledgements object

See End-to-end Acknowledgements for more information on how event acknowledgement is handled.

1 nested properties
enabled boolean | null

When enabled for a sink, any source connected to that sink, where the source supports end-to-end acknowledgements as well, waits for events to be acknowledged by all connected sinks before acknowledging them at the source.

Enabling or disabling acknowledgements at the sink level takes precedence over any global acknowledgements configuration.

assume_role string | null
auth object | object | object | object

Configuration of the authentication strategy for interacting with AWS services.

message_deduplication_id string | null

This value is a template which should result in a unique string for each event. See the AWS documentation for more about how AWS does message deduplication.

message_group_id string | null

Can be applied only to FIFO queues.

request object

Various settings can be configured, such as concurrency and rate limits, timeouts, retry behavior, etc.

Note that the retry backoff policy follows the Fibonacci sequence.

9 nested properties
adaptive_concurrency object

These parameters typically do not require changes from the default, and incorrect values can lead to meta-stable or unstable performance and sink behavior. Proceed with caution.

5 nested properties
decrease_ratio number

Valid values are greater than 0 and less than 1. Smaller values cause the algorithm to scale back rapidly when latency increases.

Note that the new limit is rounded down after applying this ratio.

Default: 0.9
min=0.0max=1.0
ewma_alpha number

Valid values are greater than 0 and less than 1.

ARC uses an exponentially weighted moving average (EWMA) of past RTT measurements as a reference to compare with the current RTT. Smaller values cause this reference to adjust more slowly, which may be useful if a service has unusually high response variability.

Default: 0.4
min=0.0max=1.0
initial_concurrency integer

It is recommended to set this value to your service's average limit if you're seeing that it takes a long time to ramp up adaptive concurrency after a restart. You can find this value by looking at the adaptive_concurrency_limit metric.

Default: 1
min=1.0max=9007199254740990.0
max_concurrency_limit integer

The adaptive request concurrency limit will not go above this bound. This is put in place as a safeguard.

Default: 200
min=1.0max=9007199254740990.0
rtt_deviation_scale number

Valid values are greater than or equal to 0, and we expect reasonable values to range from 1.0 to 3.0.

When calculating the past RTT average, we also compute a secondary “deviation” value that indicates how variable those values are. We use that deviation when comparing the past RTT average to the current measurements, so we can ignore increases in RTT that are within an expected range. This factor is used to scale up the deviation to an appropriate range. Larger values cause the algorithm to ignore larger increases in the RTT.

Default: 2.5
min=0.0max=9007199254740990.0
concurrency const: "none" | const: "adaptive" | integer

Configuration for outbound request concurrency.

This can be set either to one of the below enum values or to a positive integer, which denotes a fixed concurrency limit.

rate_limit_duration_secs integer

The time window used for the rate_limit_num option.

Default: 1
min=0.0max=9007199254740990.0
rate_limit_num integer

The maximum number of requests allowed within the rate_limit_duration_secs time window.

Default: 9223372036854775807
min=0.0max=9007199254740990.0
retry_attempts integer

The maximum number of retries to make for failed requests.

Default: 9223372036854775807
min=0.0max=9007199254740990.0
retry_initial_backoff_secs integer

After the first retry has failed, the fibonacci sequence is used to select future backoffs.

Default: 1
min=1.0max=9007199254740990.0
retry_jitter_mode const: "None" | const: "Full"

The jitter mode to use for retry backoff behavior.

retry_max_duration_secs integer

The maximum amount of time to wait between retries.

Default: 30
min=1.0max=9007199254740990.0
timeout_secs integer

Datadog highly recommends that you do not lower this value below the service's internal timeout, as this could create orphaned requests, pile on retries, and result in duplicate data downstream.

Default: 60
min=0.0max=9007199254740990.0
vector::sinks::aws_s_s::sns::config::SnsSinkConfig

Configuration for the aws_sns sink.

vector::sinks::aws_s_s::sqs::config::SqsSinkConfig

Configuration for the aws_sqs sink.

vector::sinks::axiom::AxiomConfig object

Configuration for the axiom sink.

dataset string required

The Axiom dataset to write to.

token string required

Wrapper for sensitive strings containing credentials

acknowledgements object

See End-to-end Acknowledgements for more information on how event acknowledgement is handled.

1 nested properties
enabled boolean | null

When enabled for a sink, any source connected to that sink, where the source supports end-to-end acknowledgements as well, waits for events to be acknowledged by all connected sinks before acknowledging them at the source.

Enabling or disabling acknowledgements at the sink level takes precedence over any global acknowledgements configuration.

The batch settings for the sink.

Default:
{
  "max_bytes": null,
  "max_events": null,
  "timeout_secs": null
}
compression const: "none" | const: "gzip" | const: "zlib" | const: "zstd" | const: "snappy" | object

All compression algorithms use the default compression level unless otherwise specified.

org_id string | null

Only required when using personal tokens.

request

Outbound HTTP request settings.

All of: object object, Middleware settings for outbound requests. object

Optional, constrains TLS settings for this sink.

url string | null

Only required if not using Axiom Cloud.

format=uri
vector::sinks::azure_blob::config::AzureBlobSinkConfig

Configuration for the azure_blob sink.

vector::sinks::azure_monitor_logs::config::AzureMonitorLogsConfig object

Configuration for the azure_monitor_logs sink.

customer_id string required
log_type string required

Can only contain letters, numbers, and underscores (_), and may not exceed 100 characters.

pattern=[a-zA-Z0-9_]{1,100}
shared_key string required

Wrapper for sensitive strings containing credentials

acknowledgements object

See End-to-end Acknowledgements for more information on how event acknowledgement is handled.

1 nested properties
enabled boolean | null

When enabled for a sink, any source connected to that sink, where the source supports end-to-end acknowledgements as well, waits for events to be acknowledged by all connected sinks before acknowledging them at the source.

Enabling or disabling acknowledgements at the sink level takes precedence over any global acknowledgements configuration.

azure_resource_id string | null

Event batching behavior.

Default:
{
  "max_bytes": null,
  "max_events": null,
  "timeout_secs": null
}
encoding object

Transformations to prepare an event for serialization.

3 nested properties
except_fields array | null

List of fields that are excluded from the encoded event.

Default: null
only_fields array | null

List of fields that are included in the encoded event.

Default: null
timestamp_format null | const: "unix" | const: "rfc3339" | const: "unix_ms" | const: "unix_us" | const: "unix_ns" | const: "unix_float"

Format used for timestamp fields.

Default: null
host string
Default: "ods.opinsights.azure.com"
request object

Various settings can be configured, such as concurrency and rate limits, timeouts, retry behavior, etc.

Note that the retry backoff policy follows the Fibonacci sequence.

9 nested properties
adaptive_concurrency object

These parameters typically do not require changes from the default, and incorrect values can lead to meta-stable or unstable performance and sink behavior. Proceed with caution.

5 nested properties
decrease_ratio number

Valid values are greater than 0 and less than 1. Smaller values cause the algorithm to scale back rapidly when latency increases.

Note that the new limit is rounded down after applying this ratio.

Default: 0.9
min=0.0max=1.0
ewma_alpha number

Valid values are greater than 0 and less than 1.

ARC uses an exponentially weighted moving average (EWMA) of past RTT measurements as a reference to compare with the current RTT. Smaller values cause this reference to adjust more slowly, which may be useful if a service has unusually high response variability.

Default: 0.4
min=0.0max=1.0
initial_concurrency integer

It is recommended to set this value to your service's average limit if you're seeing that it takes a long time to ramp up adaptive concurrency after a restart. You can find this value by looking at the adaptive_concurrency_limit metric.

Default: 1
min=1.0max=9007199254740990.0
max_concurrency_limit integer

The adaptive request concurrency limit will not go above this bound. This is put in place as a safeguard.

Default: 200
min=1.0max=9007199254740990.0
rtt_deviation_scale number

Valid values are greater than or equal to 0, and we expect reasonable values to range from 1.0 to 3.0.

When calculating the past RTT average, we also compute a secondary “deviation” value that indicates how variable those values are. We use that deviation when comparing the past RTT average to the current measurements, so we can ignore increases in RTT that are within an expected range. This factor is used to scale up the deviation to an appropriate range. Larger values cause the algorithm to ignore larger increases in the RTT.

Default: 2.5
min=0.0max=9007199254740990.0
concurrency const: "none" | const: "adaptive" | integer

Configuration for outbound request concurrency.

This can be set either to one of the below enum values or to a positive integer, which denotes a fixed concurrency limit.

rate_limit_duration_secs integer

The time window used for the rate_limit_num option.

Default: 1
min=0.0max=9007199254740990.0
rate_limit_num integer

The maximum number of requests allowed within the rate_limit_duration_secs time window.

Default: 9223372036854775807
min=0.0max=9007199254740990.0
retry_attempts integer

The maximum number of retries to make for failed requests.

Default: 9223372036854775807
min=0.0max=9007199254740990.0
retry_initial_backoff_secs integer

After the first retry has failed, the fibonacci sequence is used to select future backoffs.

Default: 1
min=1.0max=9007199254740990.0
retry_jitter_mode const: "None" | const: "Full"

The jitter mode to use for retry backoff behavior.

retry_max_duration_secs integer

The maximum amount of time to wait between retries.

Default: 30
min=1.0max=9007199254740990.0
timeout_secs integer

Datadog highly recommends that you do not lower this value below the service's internal timeout, as this could create orphaned requests, pile on retries, and result in duplicate data downstream.

Default: 60
min=0.0max=9007199254740990.0

The setting of log_schema.timestamp_key, usually timestamp, is used here by default. This field should be used in rare cases where TimeGenerated should point to a specific log field. For example, use this field to set the log field source_timestamp as holding the value that should be used as TimeGenerated on the Azure side.

vector::sinks::blackhole::config::BlackholeConfig object

Configuration for the blackhole sink.

acknowledgements object

See End-to-end Acknowledgements for more information on how event acknowledgement is handled.

1 nested properties
enabled boolean | null

When enabled for a sink, any source connected to that sink, where the source supports end-to-end acknowledgements as well, waits for events to be acknowledged by all connected sinks before acknowledging them at the source.

Enabling or disabling acknowledgements at the sink level takes precedence over any global acknowledgements configuration.

print_interval_secs integer

A span of time, in whole seconds.

min=0.0max=9007199254740990.0
rate integer | null

By default, there is no limit.

min=0.0max=9007199254740990.0
vector::sinks::clickhouse::config::ClickhouseConfig object

Configuration for the clickhouse sink.

endpoint string required
table string required

In many cases, components can be configured so that part of the component's functionality can be customized on a per-event basis. For example, you have a sink that writes events to a file and you want to specify which file an event should go to by using an event field as part of the input to the filename used.

By using Template, users can specify either fixed strings or templated strings. Templated strings use a common syntax to refer to fields in an event that is used as the input data when rendering the template. An example of a fixed string is my-file.log. An example of a template string is my-file-{{key}}.log, where {{key}} is the key's value when the template is rendered into a string.

acknowledgements object

See End-to-end Acknowledgements for more information on how event acknowledgement is handled.

1 nested properties
enabled boolean | null

When enabled for a sink, any source connected to that sink, where the source supports end-to-end acknowledgements as well, waits for events to be acknowledged by all connected sinks before acknowledging them at the source.

Enabling or disabling acknowledgements at the sink level takes precedence over any global acknowledgements configuration.

Event batching behavior.

Default:
{
  "max_bytes": null,
  "max_events": null,
  "timeout_secs": null
}
compression const: "none" | const: "gzip" | const: "zlib" | const: "zstd" | const: "snappy" | object

All compression algorithms use the default compression level unless otherwise specified.

The database that contains the table that data is inserted into.

date_time_best_effort boolean

Sets date_time_input_format to best_effort, allowing ClickHouse to properly parse RFC3339/ISO 8601.

Default: false
encoding object

Transformations to prepare an event for serialization.

3 nested properties
except_fields array | null

List of fields that are excluded from the encoded event.

Default: null
only_fields array | null

List of fields that are included in the encoded event.

Default: null
timestamp_format null | const: "unix" | const: "rfc3339" | const: "unix_ms" | const: "unix_us" | const: "unix_ns" | const: "unix_float"

Format used for timestamp fields.

Default: null
format const: "json_each_row" | const: "json_as_object" | const: "json_as_string"

The format to parse input data.

Default: "json_each_row"
insert_random_shard boolean

Sets insert_distributed_one_random_shard, allowing ClickHouse to insert data into a random shard when using Distributed Table Engine.

Default: false
request object

Various settings can be configured, such as concurrency and rate limits, timeouts, retry behavior, etc.

Note that the retry backoff policy follows the Fibonacci sequence.

9 nested properties
adaptive_concurrency object

These parameters typically do not require changes from the default, and incorrect values can lead to meta-stable or unstable performance and sink behavior. Proceed with caution.

5 nested properties
decrease_ratio number

Valid values are greater than 0 and less than 1. Smaller values cause the algorithm to scale back rapidly when latency increases.

Note that the new limit is rounded down after applying this ratio.

Default: 0.9
min=0.0max=1.0
ewma_alpha number

Valid values are greater than 0 and less than 1.

ARC uses an exponentially weighted moving average (EWMA) of past RTT measurements as a reference to compare with the current RTT. Smaller values cause this reference to adjust more slowly, which may be useful if a service has unusually high response variability.

Default: 0.4
min=0.0max=1.0
initial_concurrency integer

It is recommended to set this value to your service's average limit if you're seeing that it takes a long time to ramp up adaptive concurrency after a restart. You can find this value by looking at the adaptive_concurrency_limit metric.

Default: 1
min=1.0max=9007199254740990.0
max_concurrency_limit integer

The adaptive request concurrency limit will not go above this bound. This is put in place as a safeguard.

Default: 200
min=1.0max=9007199254740990.0
rtt_deviation_scale number

Valid values are greater than or equal to 0, and we expect reasonable values to range from 1.0 to 3.0.

When calculating the past RTT average, we also compute a secondary “deviation” value that indicates how variable those values are. We use that deviation when comparing the past RTT average to the current measurements, so we can ignore increases in RTT that are within an expected range. This factor is used to scale up the deviation to an appropriate range. Larger values cause the algorithm to ignore larger increases in the RTT.

Default: 2.5
min=0.0max=9007199254740990.0
concurrency const: "none" | const: "adaptive" | integer

Configuration for outbound request concurrency.

This can be set either to one of the below enum values or to a positive integer, which denotes a fixed concurrency limit.

rate_limit_duration_secs integer

The time window used for the rate_limit_num option.

Default: 1
min=0.0max=9007199254740990.0
rate_limit_num integer

The maximum number of requests allowed within the rate_limit_duration_secs time window.

Default: 9223372036854775807
min=0.0max=9007199254740990.0
retry_attempts integer

The maximum number of retries to make for failed requests.

Default: 9223372036854775807
min=0.0max=9007199254740990.0
retry_initial_backoff_secs integer

After the first retry has failed, the fibonacci sequence is used to select future backoffs.

Default: 1
min=1.0max=9007199254740990.0
retry_jitter_mode const: "None" | const: "Full"

The jitter mode to use for retry backoff behavior.

retry_max_duration_secs integer

The maximum amount of time to wait between retries.

Default: 30
min=1.0max=9007199254740990.0
timeout_secs integer

Datadog highly recommends that you do not lower this value below the service's internal timeout, as this could create orphaned requests, pile on retries, and result in duplicate data downstream.

Default: 60
min=0.0max=9007199254740990.0
skip_unknown_fields boolean | null

If left unspecified, use the default provided by the ClickHouse server.

Default: null
vector::sinks::console::config::ConsoleSinkConfig

Configuration for the console sink.

vector::sinks::databend::config::DatabendConfig object

Configuration for the databend sink.

endpoint string required
table string required

The table that data is inserted into.

acknowledgements object

See End-to-end Acknowledgements for more information on how event acknowledgement is handled.

1 nested properties
enabled boolean | null

When enabled for a sink, any source connected to that sink, where the source supports end-to-end acknowledgements as well, waits for events to be acknowledged by all connected sinks before acknowledging them at the source.

Enabling or disabling acknowledgements at the sink level takes precedence over any global acknowledgements configuration.

The username and password to authenticate with. Overrides the username and password in DSN.

Event batching behavior.

Default:
{
  "max_bytes": null,
  "max_events": null,
  "timeout_secs": null
}
compression const: "none" | const: "gzip"

Compression configuration.

Default: "none"
database string | null

The database that contains the table that data is inserted into. Overrides the database in DSN.

encoding

Configures how events are encoded into raw bytes.

Default:
{
  "codec": "json",
  "json": {
    "pretty": false
  }
}
All of: variant, object object
missing_field_as const: "ERROR" | const: "NULL" | const: "FIELD_DEFAULT" | const: "TYPE_DEFAULT"

Defines how missing fields are handled for NDJson. Refer to https://docs.databend.com/sql/sql-reference/file-format-options#null_field_as

Default: "NULL"
request object

Various settings can be configured, such as concurrency and rate limits, timeouts, retry behavior, etc.

Note that the retry backoff policy follows the Fibonacci sequence.

9 nested properties
adaptive_concurrency object

These parameters typically do not require changes from the default, and incorrect values can lead to meta-stable or unstable performance and sink behavior. Proceed with caution.

5 nested properties
decrease_ratio number

Valid values are greater than 0 and less than 1. Smaller values cause the algorithm to scale back rapidly when latency increases.

Note that the new limit is rounded down after applying this ratio.

Default: 0.9
min=0.0max=1.0
ewma_alpha number

Valid values are greater than 0 and less than 1.

ARC uses an exponentially weighted moving average (EWMA) of past RTT measurements as a reference to compare with the current RTT. Smaller values cause this reference to adjust more slowly, which may be useful if a service has unusually high response variability.

Default: 0.4
min=0.0max=1.0
initial_concurrency integer

It is recommended to set this value to your service's average limit if you're seeing that it takes a long time to ramp up adaptive concurrency after a restart. You can find this value by looking at the adaptive_concurrency_limit metric.

Default: 1
min=1.0max=9007199254740990.0
max_concurrency_limit integer

The adaptive request concurrency limit will not go above this bound. This is put in place as a safeguard.

Default: 200
min=1.0max=9007199254740990.0
rtt_deviation_scale number

Valid values are greater than or equal to 0, and we expect reasonable values to range from 1.0 to 3.0.

When calculating the past RTT average, we also compute a secondary “deviation” value that indicates how variable those values are. We use that deviation when comparing the past RTT average to the current measurements, so we can ignore increases in RTT that are within an expected range. This factor is used to scale up the deviation to an appropriate range. Larger values cause the algorithm to ignore larger increases in the RTT.

Default: 2.5
min=0.0max=9007199254740990.0
concurrency const: "none" | const: "adaptive" | integer

Configuration for outbound request concurrency.

This can be set either to one of the below enum values or to a positive integer, which denotes a fixed concurrency limit.

rate_limit_duration_secs integer

The time window used for the rate_limit_num option.

Default: 1
min=0.0max=9007199254740990.0
rate_limit_num integer

The maximum number of requests allowed within the rate_limit_duration_secs time window.

Default: 9223372036854775807
min=0.0max=9007199254740990.0
retry_attempts integer

The maximum number of retries to make for failed requests.

Default: 9223372036854775807
min=0.0max=9007199254740990.0
retry_initial_backoff_secs integer

After the first retry has failed, the fibonacci sequence is used to select future backoffs.

Default: 1
min=1.0max=9007199254740990.0
retry_jitter_mode const: "None" | const: "Full"

The jitter mode to use for retry backoff behavior.

retry_max_duration_secs integer

The maximum amount of time to wait between retries.

Default: 30
min=1.0max=9007199254740990.0
timeout_secs integer

Datadog highly recommends that you do not lower this value below the service's internal timeout, as this could create orphaned requests, pile on retries, and result in duplicate data downstream.

Default: 60
min=0.0max=9007199254740990.0

The TLS configuration to use when connecting to the Databend server.

vector::sinks::datadog::LocalDatadogCommonConfig object

Shared configuration for Datadog sinks. Contains the maximum set of common settings that applies to all DD sink components.

acknowledgements object

See End-to-end Acknowledgements for more information on how event acknowledgement is handled.

1 nested properties
enabled boolean | null

When enabled for a sink, any source connected to that sink, where the source supports end-to-end acknowledgements as well, waits for events to be acknowledged by all connected sinks before acknowledging them at the source.

Enabling or disabling acknowledgements at the sink level takes precedence over any global acknowledgements configuration.

If an event has a Datadog API key set explicitly in its metadata, it takes precedence over this setting.

This value can also be set by specifying the DD_API_KEY environment variable. The value specified here takes precedence over the environment variable.

endpoint string | null

The endpoint must contain an HTTP scheme, and may specify a hostname or IP address and port. The API path should NOT be specified as this is handled by the sink.

If set, overrides the site option.

Default: null
site string | null

This value can also be set by specifying the DD_SITE environment variable. The value specified here takes precedence over the environment variable.

If not specified by the environment variable, a default value of datadoghq.com is taken.

vector::sinks::datadog::events::config::DatadogEventsConfig

Configuration for the datadog_events sink.

vector::sinks::datadog::logs::config::DatadogLogsConfig

Configuration for the datadog_logs sink.

vector::sinks::datadog::metrics::config::DatadogMetricsConfig

Configuration for the datadog_metrics sink.

vector::sinks::datadog::traces::config::DatadogTracesConfig

Configuration for the datadog_traces sink.

vector::sinks::elasticsearch::config::ElasticsearchConfig object

Configuration for the elasticsearch sink.

acknowledgements object

See End-to-end Acknowledgements for more information on how event acknowledgement is handled.

1 nested properties
enabled boolean | null

When enabled for a sink, any source connected to that sink, where the source supports end-to-end acknowledgements as well, waits for events to be acknowledged by all connected sinks before acknowledging them at the source.

Enabling or disabling acknowledgements at the sink level takes precedence over any global acknowledgements configuration.

api_version const: "auto" | const: "v6" | const: "v7" | const: "v8"

Amazon OpenSearch Serverless requires this option to be set to auto (the default).

Default: "auto"
auth null | object

Elasticsearch Authentication strategies.

Event batching behavior.

Default:
{
  "max_bytes": null,
  "max_events": null,
  "timeout_secs": null
}
bulk object

Elasticsearch bulk mode configuration.

Default:
{
  "action": "index",
  "index": "vector-%Y.%m.%d",
  "template_fallback_index": null,
  "version": null,
  "version_type": "internal"
}
5 nested properties
action string

In many cases, components can be configured so that part of the component's functionality can be customized on a per-event basis. For example, you have a sink that writes events to a file and you want to specify which file an event should go to by using an event field as part of the input to the filename used.

By using Template, users can specify either fixed strings or templated strings. Templated strings use a common syntax to refer to fields in an event that is used as the input data when rendering the template. An example of a fixed string is my-file.log. An example of a template string is my-file-{{key}}.log, where {{key}} is the key's value when the template is rendered into a string.

index string

In many cases, components can be configured so that part of the component's functionality can be customized on a per-event basis. For example, you have a sink that writes events to a file and you want to specify which file an event should go to by using an event field as part of the input to the filename used.

By using Template, users can specify either fixed strings or templated strings. Templated strings use a common syntax to refer to fields in an event that is used as the input data when rendering the template. An example of a fixed string is my-file.log. An example of a template string is my-file-{{key}}.log, where {{key}} is the key's value when the template is rendered into a string.

template_fallback_index string | null

The default index to write events to if the template in bulk.index cannot be resolved

version_type const: "internal" | const: "external" | const: "external_gte"

Possible values are internal, external or external_gt and external_gte.

Default: "internal"
compression const: "none" | const: "gzip" | const: "zlib" | const: "zstd" | const: "snappy" | object

All compression algorithms use the default compression level unless otherwise specified.

data_stream null | object

Elasticsearch data stream mode configuration.

Default: null
distribution null | object

Options for determining the health of an endpoint.

Default: null
doc_type string

This is only relevant for Elasticsearch <= 6.X. If you are using >= 7.0 you do not need to set this option since Elasticsearch has removed it.

Default: "_doc"
encoding object

Transformations to prepare an event for serialization.

3 nested properties
except_fields array | null

List of fields that are excluded from the encoded event.

Default: null
only_fields array | null

List of fields that are included in the encoded event.

Default: null
timestamp_format null | const: "unix" | const: "rfc3339" | const: "unix_ms" | const: "unix_us" | const: "unix_ns" | const: "unix_float"

Format used for timestamp fields.

Default: null
endpoint string | null

The endpoint must contain an HTTP scheme, and may specify a hostname or IP address and port.

Default: null
endpoints string[]

The endpoint must contain an HTTP scheme, and may specify a hostname or IP address and port.

Default:
[]

By default, the _id field is not set, which allows Elasticsearch to set this automatically. Setting your own Elasticsearch IDs can hinder performance.

Default: null

Configuration for the metric_to_log transform.

Default: null
mode const: "bulk" | const: "data_stream"

Elasticsearch Indexing mode.

Default: "bulk"
opensearch_service_type const: "managed" | const: "serverless"

Amazon OpenSearch service type

Default: "managed"
pipeline string | null

The name of the pipeline to apply.

Default: null
query Record<string, string>

Custom parameters to add to the query string for each HTTP request sent to Elasticsearch.

Default: null
request

Outbound HTTP request settings.

All of: object object, Middleware settings for outbound requests. object
request_retry_partial boolean

To avoid duplicates in Elasticsearch, please use option id_key.

Default: false
suppress_type_name boolean

The type field was deprecated in Elasticsearch 7.x and removed in Elasticsearch 8.x.

If enabled, the doc_type option is ignored.

Default: false
vector::sinks::file::FileSinkConfig

Configuration for the file sink.

vector::sinks::gcp::cloud_storage::GcsSinkConfig

Configuration for the gcp_cloud_storage sink.

vector::sinks::gcp::pubsub::PubsubConfig

Configuration for the gcp_pubsub sink.

vector::sinks::gcp::stackdriver::logs::config::StackdriverConfig

Configuration for the gcp_stackdriver_logs sink.

vector::sinks::gcp::stackdriver::metrics::config::StackdriverConfig

Configuration for the gcp_stackdriver_metrics sink.

vector::sinks::gcp_chronicle::chronicle_unstructured::ChronicleUnstructuredConfig

Configuration for the gcp_chronicle_unstructured sink.

vector::sinks::greptimedb::logs::config::GreptimeDBLogsConfig object

Configuration for the greptimedb_logs sink.

endpoint string required

The endpoint of the GreptimeDB server.

table string required

In many cases, components can be configured so that part of the component's functionality can be customized on a per-event basis. For example, you have a sink that writes events to a file and you want to specify which file an event should go to by using an event field as part of the input to the filename used.

By using Template, users can specify either fixed strings or templated strings. Templated strings use a common syntax to refer to fields in an event that is used as the input data when rendering the template. An example of a fixed string is my-file.log. An example of a template string is my-file-{{key}}.log, where {{key}} is the key's value when the template is rendered into a string.

acknowledgements object

See End-to-end Acknowledgements for more information on how event acknowledgement is handled.

1 nested properties
enabled boolean | null

When enabled for a sink, any source connected to that sink, where the source supports end-to-end acknowledgements as well, waits for events to be acknowledged by all connected sinks before acknowledging them at the source.

Enabling or disabling acknowledgements at the sink level takes precedence over any global acknowledgements configuration.

Event batching behavior.

Default:
{
  "max_bytes": null,
  "max_events": null,
  "timeout_secs": null
}
compression const: "none" | const: "gzip" | const: "zlib" | const: "zstd" | const: "snappy" | object

All compression algorithms use the default compression level unless otherwise specified.

dbname string

In many cases, components can be configured so that part of the component's functionality can be customized on a per-event basis. For example, you have a sink that writes events to a file and you want to specify which file an event should go to by using an event field as part of the input to the filename used.

By using Template, users can specify either fixed strings or templated strings. Templated strings use a common syntax to refer to fields in an event that is used as the input data when rendering the template. An example of a fixed string is my-file.log. An example of a template string is my-file-{{key}}.log, where {{key}} is the key's value when the template is rendered into a string.

encoding object

Transformations to prepare an event for serialization.

3 nested properties
except_fields array | null

List of fields that are excluded from the encoded event.

Default: null
only_fields array | null

List of fields that are included in the encoded event.

Default: null
timestamp_format null | const: "unix" | const: "rfc3339" | const: "unix_ms" | const: "unix_us" | const: "unix_ns" | const: "unix_float"

Format used for timestamp fields.

Default: null
extra_params Record<string, string>

Custom parameters to add to the query string for each HTTP request sent to GreptimeDB.

Default: null

This is required if your instance has authentication enabled.

Default: null
pipeline_name string

In many cases, components can be configured so that part of the component's functionality can be customized on a per-event basis. For example, you have a sink that writes events to a file and you want to specify which file an event should go to by using an event field as part of the input to the filename used.

By using Template, users can specify either fixed strings or templated strings. Templated strings use a common syntax to refer to fields in an event that is used as the input data when rendering the template. An example of a fixed string is my-file.log. An example of a template string is my-file-{{key}}.log, where {{key}} is the key's value when the template is rendered into a string.

Pipeline version to be used for the logs.

request object

Various settings can be configured, such as concurrency and rate limits, timeouts, retry behavior, etc.

Note that the retry backoff policy follows the Fibonacci sequence.

9 nested properties
adaptive_concurrency object

These parameters typically do not require changes from the default, and incorrect values can lead to meta-stable or unstable performance and sink behavior. Proceed with caution.

5 nested properties
decrease_ratio number

Valid values are greater than 0 and less than 1. Smaller values cause the algorithm to scale back rapidly when latency increases.

Note that the new limit is rounded down after applying this ratio.

Default: 0.9
min=0.0max=1.0
ewma_alpha number

Valid values are greater than 0 and less than 1.

ARC uses an exponentially weighted moving average (EWMA) of past RTT measurements as a reference to compare with the current RTT. Smaller values cause this reference to adjust more slowly, which may be useful if a service has unusually high response variability.

Default: 0.4
min=0.0max=1.0
initial_concurrency integer

It is recommended to set this value to your service's average limit if you're seeing that it takes a long time to ramp up adaptive concurrency after a restart. You can find this value by looking at the adaptive_concurrency_limit metric.

Default: 1
min=1.0max=9007199254740990.0
max_concurrency_limit integer

The adaptive request concurrency limit will not go above this bound. This is put in place as a safeguard.

Default: 200
min=1.0max=9007199254740990.0
rtt_deviation_scale number

Valid values are greater than or equal to 0, and we expect reasonable values to range from 1.0 to 3.0.

When calculating the past RTT average, we also compute a secondary “deviation” value that indicates how variable those values are. We use that deviation when comparing the past RTT average to the current measurements, so we can ignore increases in RTT that are within an expected range. This factor is used to scale up the deviation to an appropriate range. Larger values cause the algorithm to ignore larger increases in the RTT.

Default: 2.5
min=0.0max=9007199254740990.0
concurrency const: "none" | const: "adaptive" | integer

Configuration for outbound request concurrency.

This can be set either to one of the below enum values or to a positive integer, which denotes a fixed concurrency limit.

rate_limit_duration_secs integer

The time window used for the rate_limit_num option.

Default: 1
min=0.0max=9007199254740990.0
rate_limit_num integer

The maximum number of requests allowed within the rate_limit_duration_secs time window.

Default: 9223372036854775807
min=0.0max=9007199254740990.0
retry_attempts integer

The maximum number of retries to make for failed requests.

Default: 9223372036854775807
min=0.0max=9007199254740990.0
retry_initial_backoff_secs integer

After the first retry has failed, the fibonacci sequence is used to select future backoffs.

Default: 1
min=1.0max=9007199254740990.0
retry_jitter_mode const: "None" | const: "Full"

The jitter mode to use for retry backoff behavior.

retry_max_duration_secs integer

The maximum amount of time to wait between retries.

Default: 30
min=1.0max=9007199254740990.0
timeout_secs integer

Datadog highly recommends that you do not lower this value below the service's internal timeout, as this could create orphaned requests, pile on retries, and result in duplicate data downstream.

Default: 60
min=0.0max=9007199254740990.0
username string | null

This is required if your instance has authentication enabled.

Default: null
vector::sinks::greptimedb::metrics::config::GreptimeDBConfig object

Configuration items for GreptimeDB

endpoint string required

This sink uses GreptimeDB's gRPC interface for data ingestion. By default, GreptimeDB listens to port 4001 for gRPC protocol.

The address must include a port.

acknowledgements object

See End-to-end Acknowledgements for more information on how event acknowledgement is handled.

1 nested properties
enabled boolean | null

When enabled for a sink, any source connected to that sink, where the source supports end-to-end acknowledgements as well, waits for events to be acknowledged by all connected sinks before acknowledging them at the source.

Enabling or disabling acknowledgements at the sink level takes precedence over any global acknowledgements configuration.

Event batching behavior.

Default:
{
  "max_bytes": null,
  "max_events": null,
  "timeout_secs": null
}
dbname string

Default to public, the default database of GreptimeDB.

Database can be created via create database statement on GreptimeDB. If you are using GreptimeCloud, use dbname from the connection information of your instance.

Default: "public"
grpc_compression string | null

Set gRPC compression encoding for the request Default to none, gzip or zstd is supported.

Default: null
new_naming boolean | null

This is to keep consistency with GreptimeDB's naming pattern. By default, this sink will use val for value column name, and ts for time index name. When turned on, greptime_value and greptime_timestamp will be used for these names.

If you are using this Vector sink together with other data ingestion sources of GreptimeDB, like Prometheus Remote Write and Influxdb Line Protocol, it is highly recommended to turn on this.

Also if there is a tag name conflict from your data source, for example, you have a tag named as val or ts, you need to turn on this option to avoid the conflict.

Default to false for compatibility.

This is required if your instance has authentication enabled.

Default: null
request object

Various settings can be configured, such as concurrency and rate limits, timeouts, retry behavior, etc.

Note that the retry backoff policy follows the Fibonacci sequence.

9 nested properties
adaptive_concurrency object

These parameters typically do not require changes from the default, and incorrect values can lead to meta-stable or unstable performance and sink behavior. Proceed with caution.

5 nested properties
decrease_ratio number

Valid values are greater than 0 and less than 1. Smaller values cause the algorithm to scale back rapidly when latency increases.

Note that the new limit is rounded down after applying this ratio.

Default: 0.9
min=0.0max=1.0
ewma_alpha number

Valid values are greater than 0 and less than 1.

ARC uses an exponentially weighted moving average (EWMA) of past RTT measurements as a reference to compare with the current RTT. Smaller values cause this reference to adjust more slowly, which may be useful if a service has unusually high response variability.

Default: 0.4
min=0.0max=1.0
initial_concurrency integer

It is recommended to set this value to your service's average limit if you're seeing that it takes a long time to ramp up adaptive concurrency after a restart. You can find this value by looking at the adaptive_concurrency_limit metric.

Default: 1
min=1.0max=9007199254740990.0
max_concurrency_limit integer

The adaptive request concurrency limit will not go above this bound. This is put in place as a safeguard.

Default: 200
min=1.0max=9007199254740990.0
rtt_deviation_scale number

Valid values are greater than or equal to 0, and we expect reasonable values to range from 1.0 to 3.0.

When calculating the past RTT average, we also compute a secondary “deviation” value that indicates how variable those values are. We use that deviation when comparing the past RTT average to the current measurements, so we can ignore increases in RTT that are within an expected range. This factor is used to scale up the deviation to an appropriate range. Larger values cause the algorithm to ignore larger increases in the RTT.

Default: 2.5
min=0.0max=9007199254740990.0
concurrency const: "none" | const: "adaptive" | integer

Configuration for outbound request concurrency.

This can be set either to one of the below enum values or to a positive integer, which denotes a fixed concurrency limit.

rate_limit_duration_secs integer

The time window used for the rate_limit_num option.

Default: 1
min=0.0max=9007199254740990.0
rate_limit_num integer

The maximum number of requests allowed within the rate_limit_duration_secs time window.

Default: 9223372036854775807
min=0.0max=9007199254740990.0
retry_attempts integer

The maximum number of retries to make for failed requests.

Default: 9223372036854775807
min=0.0max=9007199254740990.0
retry_initial_backoff_secs integer

After the first retry has failed, the fibonacci sequence is used to select future backoffs.

Default: 1
min=1.0max=9007199254740990.0
retry_jitter_mode const: "None" | const: "Full"

The jitter mode to use for retry backoff behavior.

retry_max_duration_secs integer

The maximum amount of time to wait between retries.

Default: 30
min=1.0max=9007199254740990.0
timeout_secs integer

Datadog highly recommends that you do not lower this value below the service's internal timeout, as this could create orphaned requests, pile on retries, and result in duplicate data downstream.

Default: 60
min=0.0max=9007199254740990.0
username string | null

This is required if your instance has authentication enabled.

Default: null
vector::sinks::greptimedb::metrics::config::GreptimeDBMetricsConfig object

Configuration items for GreptimeDB

endpoint string required

This sink uses GreptimeDB's gRPC interface for data ingestion. By default, GreptimeDB listens to port 4001 for gRPC protocol.

The address must include a port.

acknowledgements object

See End-to-end Acknowledgements for more information on how event acknowledgement is handled.

1 nested properties
enabled boolean | null

When enabled for a sink, any source connected to that sink, where the source supports end-to-end acknowledgements as well, waits for events to be acknowledged by all connected sinks before acknowledging them at the source.

Enabling or disabling acknowledgements at the sink level takes precedence over any global acknowledgements configuration.

Event batching behavior.

Default:
{
  "max_bytes": null,
  "max_events": null,
  "timeout_secs": null
}
dbname string

Default to public, the default database of GreptimeDB.

Database can be created via create database statement on GreptimeDB. If you are using GreptimeCloud, use dbname from the connection information of your instance.

Default: "public"
grpc_compression string | null

Set gRPC compression encoding for the request Default to none, gzip or zstd is supported.

Default: null
new_naming boolean | null

This is to keep consistency with GreptimeDB's naming pattern. By default, this sink will use val for value column name, and ts for time index name. When turned on, greptime_value and greptime_timestamp will be used for these names.

If you are using this Vector sink together with other data ingestion sources of GreptimeDB, like Prometheus Remote Write and Influxdb Line Protocol, it is highly recommended to turn on this.

Also if there is a tag name conflict from your data source, for example, you have a tag named as val or ts, you need to turn on this option to avoid the conflict.

Default to false for compatibility.

This is required if your instance has authentication enabled.

Default: null
request object

Various settings can be configured, such as concurrency and rate limits, timeouts, retry behavior, etc.

Note that the retry backoff policy follows the Fibonacci sequence.

9 nested properties
adaptive_concurrency object

These parameters typically do not require changes from the default, and incorrect values can lead to meta-stable or unstable performance and sink behavior. Proceed with caution.

5 nested properties
decrease_ratio number

Valid values are greater than 0 and less than 1. Smaller values cause the algorithm to scale back rapidly when latency increases.

Note that the new limit is rounded down after applying this ratio.

Default: 0.9
min=0.0max=1.0
ewma_alpha number

Valid values are greater than 0 and less than 1.

ARC uses an exponentially weighted moving average (EWMA) of past RTT measurements as a reference to compare with the current RTT. Smaller values cause this reference to adjust more slowly, which may be useful if a service has unusually high response variability.

Default: 0.4
min=0.0max=1.0
initial_concurrency integer

It is recommended to set this value to your service's average limit if you're seeing that it takes a long time to ramp up adaptive concurrency after a restart. You can find this value by looking at the adaptive_concurrency_limit metric.

Default: 1
min=1.0max=9007199254740990.0
max_concurrency_limit integer

The adaptive request concurrency limit will not go above this bound. This is put in place as a safeguard.

Default: 200
min=1.0max=9007199254740990.0
rtt_deviation_scale number

Valid values are greater than or equal to 0, and we expect reasonable values to range from 1.0 to 3.0.

When calculating the past RTT average, we also compute a secondary “deviation” value that indicates how variable those values are. We use that deviation when comparing the past RTT average to the current measurements, so we can ignore increases in RTT that are within an expected range. This factor is used to scale up the deviation to an appropriate range. Larger values cause the algorithm to ignore larger increases in the RTT.

Default: 2.5
min=0.0max=9007199254740990.0
concurrency const: "none" | const: "adaptive" | integer

Configuration for outbound request concurrency.

This can be set either to one of the below enum values or to a positive integer, which denotes a fixed concurrency limit.

rate_limit_duration_secs integer

The time window used for the rate_limit_num option.

Default: 1
min=0.0max=9007199254740990.0
rate_limit_num integer

The maximum number of requests allowed within the rate_limit_duration_secs time window.

Default: 9223372036854775807
min=0.0max=9007199254740990.0
retry_attempts integer

The maximum number of retries to make for failed requests.

Default: 9223372036854775807
min=0.0max=9007199254740990.0
retry_initial_backoff_secs integer

After the first retry has failed, the fibonacci sequence is used to select future backoffs.

Default: 1
min=1.0max=9007199254740990.0
retry_jitter_mode const: "None" | const: "Full"

The jitter mode to use for retry backoff behavior.

retry_max_duration_secs integer

The maximum amount of time to wait between retries.

Default: 30
min=1.0max=9007199254740990.0
timeout_secs integer

Datadog highly recommends that you do not lower this value below the service's internal timeout, as this could create orphaned requests, pile on retries, and result in duplicate data downstream.

Default: 60
min=0.0max=9007199254740990.0
username string | null

This is required if your instance has authentication enabled.

Default: null
vector::sinks::honeycomb::config::HoneycombConfig object

Configuration for the honeycomb sink.

api_key string required

Wrapper for sensitive strings containing credentials

dataset string required

The dataset to which logs are sent.

acknowledgements object

See End-to-end Acknowledgements for more information on how event acknowledgement is handled.

1 nested properties
enabled boolean | null

When enabled for a sink, any source connected to that sink, where the source supports end-to-end acknowledgements as well, waits for events to be acknowledged by all connected sinks before acknowledging them at the source.

Enabling or disabling acknowledgements at the sink level takes precedence over any global acknowledgements configuration.

batch object

Event batching behavior.

Default:
{
  "max_bytes": null,
  "max_events": null,
  "timeout_secs": null
}
3 nested properties
max_bytes integer | null

This is based on the uncompressed size of the batched events, before they are serialized/compressed.

Default: 100000
min=0.0max=9007199254740990.0
max_events integer | null

The maximum size of a batch before it is flushed.

Default: null
min=0.0max=9007199254740990.0
timeout_secs number | null

The maximum age of a batch before it is flushed.

Default: 1.0
min=-9007199254740990.0max=9007199254740990.0
compression const: "none" | const: "gzip" | const: "zlib" | const: "zstd" | const: "snappy" | object

All compression algorithms use the default compression level unless otherwise specified.

encoding object

Transformations to prepare an event for serialization.

3 nested properties
except_fields array | null

List of fields that are excluded from the encoded event.

Default: null
only_fields array | null

List of fields that are included in the encoded event.

Default: null
timestamp_format null | const: "unix" | const: "rfc3339" | const: "unix_ms" | const: "unix_us" | const: "unix_ns" | const: "unix_float"

Format used for timestamp fields.

Default: null
endpoint string

Honeycomb's endpoint to send logs to

Default: "https://api.honeycomb.io"
format=uri
request object

Various settings can be configured, such as concurrency and rate limits, timeouts, retry behavior, etc.

Note that the retry backoff policy follows the Fibonacci sequence.

9 nested properties
adaptive_concurrency object

These parameters typically do not require changes from the default, and incorrect values can lead to meta-stable or unstable performance and sink behavior. Proceed with caution.

5 nested properties
decrease_ratio number

Valid values are greater than 0 and less than 1. Smaller values cause the algorithm to scale back rapidly when latency increases.

Note that the new limit is rounded down after applying this ratio.

Default: 0.9
min=0.0max=1.0
ewma_alpha number

Valid values are greater than 0 and less than 1.

ARC uses an exponentially weighted moving average (EWMA) of past RTT measurements as a reference to compare with the current RTT. Smaller values cause this reference to adjust more slowly, which may be useful if a service has unusually high response variability.

Default: 0.4
min=0.0max=1.0
initial_concurrency integer

It is recommended to set this value to your service's average limit if you're seeing that it takes a long time to ramp up adaptive concurrency after a restart. You can find this value by looking at the adaptive_concurrency_limit metric.

Default: 1
min=1.0max=9007199254740990.0
max_concurrency_limit integer

The adaptive request concurrency limit will not go above this bound. This is put in place as a safeguard.

Default: 200
min=1.0max=9007199254740990.0
rtt_deviation_scale number

Valid values are greater than or equal to 0, and we expect reasonable values to range from 1.0 to 3.0.

When calculating the past RTT average, we also compute a secondary “deviation” value that indicates how variable those values are. We use that deviation when comparing the past RTT average to the current measurements, so we can ignore increases in RTT that are within an expected range. This factor is used to scale up the deviation to an appropriate range. Larger values cause the algorithm to ignore larger increases in the RTT.

Default: 2.5
min=0.0max=9007199254740990.0
concurrency const: "none" | const: "adaptive" | integer

Configuration for outbound request concurrency.

This can be set either to one of the below enum values or to a positive integer, which denotes a fixed concurrency limit.

rate_limit_duration_secs integer

The time window used for the rate_limit_num option.

Default: 1
min=0.0max=9007199254740990.0
rate_limit_num integer

The maximum number of requests allowed within the rate_limit_duration_secs time window.

Default: 9223372036854775807
min=0.0max=9007199254740990.0
retry_attempts integer

The maximum number of retries to make for failed requests.

Default: 9223372036854775807
min=0.0max=9007199254740990.0
retry_initial_backoff_secs integer

After the first retry has failed, the fibonacci sequence is used to select future backoffs.

Default: 1
min=1.0max=9007199254740990.0
retry_jitter_mode const: "None" | const: "Full"

The jitter mode to use for retry backoff behavior.

retry_max_duration_secs integer

The maximum amount of time to wait between retries.

Default: 30
min=1.0max=9007199254740990.0
timeout_secs integer

Datadog highly recommends that you do not lower this value below the service's internal timeout, as this could create orphaned requests, pile on retries, and result in duplicate data downstream.

Default: 60
min=0.0max=9007199254740990.0
vector::sinks::http::config::HttpSinkConfig

Configuration for the http sink.

vector::sinks::humio::logs::HumioLogsConfig object

Configuration for the humio_logs sink.

encoding required

Configures how events are encoded into raw bytes.

All of: object | object | object | object | object | object object | object | object | object | object | object, object object
token string required

Wrapper for sensitive strings containing credentials

acknowledgements object

See End-to-end Acknowledgements for more information on how event acknowledgement is handled.

1 nested properties
enabled boolean | null

When enabled for a sink, any source connected to that sink, where the source supports end-to-end acknowledgements as well, waits for events to be acknowledged by all connected sinks before acknowledging them at the source.

Enabling or disabling acknowledgements at the sink level takes precedence over any global acknowledgements configuration.

Event batching behavior.

Default:
{
  "max_bytes": null,
  "max_events": null,
  "timeout_secs": null
}
compression const: "none" | const: "gzip" | const: "zlib" | const: "zstd" | const: "snappy" | object

All compression algorithms use the default compression level unless otherwise specified.

endpoint string

The scheme (http or https) must be specified. No path should be included since the paths defined by the Splunk API are used.

Default: "https://cloud.humio.com"

If unset, Humio defaults it to none.

host_key string

An optional path that deserializes an empty string to None.

In public-facing APIs, this must (if present) be equal to the repository used to create the ingest token used for authentication.

In private cluster setups, Humio can be configured to allow these to be different.

For more information, see Humio’s Format of Data.

Default: null

Can be used to tag events by specifying fields starting with #.

For more information, see Humio’s Format of Data.

Default:
[]
request object

Various settings can be configured, such as concurrency and rate limits, timeouts, retry behavior, etc.

Note that the retry backoff policy follows the Fibonacci sequence.

9 nested properties
adaptive_concurrency object

These parameters typically do not require changes from the default, and incorrect values can lead to meta-stable or unstable performance and sink behavior. Proceed with caution.

5 nested properties
decrease_ratio number

Valid values are greater than 0 and less than 1. Smaller values cause the algorithm to scale back rapidly when latency increases.

Note that the new limit is rounded down after applying this ratio.

Default: 0.9
min=0.0max=1.0
ewma_alpha number

Valid values are greater than 0 and less than 1.

ARC uses an exponentially weighted moving average (EWMA) of past RTT measurements as a reference to compare with the current RTT. Smaller values cause this reference to adjust more slowly, which may be useful if a service has unusually high response variability.

Default: 0.4
min=0.0max=1.0
initial_concurrency integer

It is recommended to set this value to your service's average limit if you're seeing that it takes a long time to ramp up adaptive concurrency after a restart. You can find this value by looking at the adaptive_concurrency_limit metric.

Default: 1
min=1.0max=9007199254740990.0
max_concurrency_limit integer

The adaptive request concurrency limit will not go above this bound. This is put in place as a safeguard.

Default: 200
min=1.0max=9007199254740990.0
rtt_deviation_scale number

Valid values are greater than or equal to 0, and we expect reasonable values to range from 1.0 to 3.0.

When calculating the past RTT average, we also compute a secondary “deviation” value that indicates how variable those values are. We use that deviation when comparing the past RTT average to the current measurements, so we can ignore increases in RTT that are within an expected range. This factor is used to scale up the deviation to an appropriate range. Larger values cause the algorithm to ignore larger increases in the RTT.

Default: 2.5
min=0.0max=9007199254740990.0
concurrency const: "none" | const: "adaptive" | integer

Configuration for outbound request concurrency.

This can be set either to one of the below enum values or to a positive integer, which denotes a fixed concurrency limit.

rate_limit_duration_secs integer

The time window used for the rate_limit_num option.

Default: 1
min=0.0max=9007199254740990.0
rate_limit_num integer

The maximum number of requests allowed within the rate_limit_duration_secs time window.

Default: 9223372036854775807
min=0.0max=9007199254740990.0
retry_attempts integer

The maximum number of retries to make for failed requests.

Default: 9223372036854775807
min=0.0max=9007199254740990.0
retry_initial_backoff_secs integer

After the first retry has failed, the fibonacci sequence is used to select future backoffs.

Default: 1
min=1.0max=9007199254740990.0
retry_jitter_mode const: "None" | const: "Full"

The jitter mode to use for retry backoff behavior.

retry_max_duration_secs integer

The maximum amount of time to wait between retries.

Default: 30
min=1.0max=9007199254740990.0
timeout_secs integer

Datadog highly recommends that you do not lower this value below the service's internal timeout, as this could create orphaned requests, pile on retries, and result in duplicate data downstream.

Default: 60
min=0.0max=9007199254740990.0

Typically the filename the logs originated from. Maps to @source in Humio.

timestamp_key string

An optional path that deserializes an empty string to None.

timestamp_nanos_key string | null

Overrides the name of the log field used to retrieve the nanosecond-enabled timestamp to send to Humio.

Default: "@timestamp.nanos"
vector::sinks::humio::metrics::HumioMetricsConfig

Configuration for the humio_metrics sink.

vector::sinks::influxdb::logs::InfluxDbLogsConfig

Configuration for the influxdb_logs sink.

vector::sinks::influxdb::metrics::InfluxDbConfig

Configuration for the influxdb_metrics sink.

vector::sinks::kafka::config::KafkaSinkConfig

Configuration for the kafka sink.

vector::sinks::loki::config::LokiConfig object

Configuration for the loki sink.

encoding required

Configures how events are encoded into raw bytes.

All of: object | object | object | object | object | object object | object | object | object | object | object, object object
endpoint string required
acknowledgements object

See End-to-end Acknowledgements for more information on how event acknowledgement is handled.

1 nested properties
enabled boolean | null

When enabled for a sink, any source connected to that sink, where the source supports end-to-end acknowledgements as well, waits for events to be acknowledged by all connected sinks before acknowledging them at the source.

Enabling or disabling acknowledgements at the sink level takes precedence over any global acknowledgements configuration.

batch object

Event batching behavior.

Default:
{
  "max_bytes": null,
  "max_events": null,
  "timeout_secs": null
}
3 nested properties
max_bytes integer | null

This is based on the uncompressed size of the batched events, before they are serialized/compressed.

Default: 1000000
min=0.0max=9007199254740990.0
max_events integer | null

The maximum size of a batch before it is flushed.

Default: 100000
min=0.0max=9007199254740990.0
timeout_secs number | null

The maximum age of a batch before it is flushed.

Default: 1.0
min=-9007199254740990.0max=9007199254740990.0
compression const: "none" | const: "gzip" | const: "zlib" | const: "zstd" | const: "snappy" | object

All compression algorithms use the default compression level unless otherwise specified.

labels Record<string, string>

Both keys and values are templateable, which enables you to attach dynamic labels to events.

Valid label keys include *, and prefixes ending with *, to allow for the expansion of objects into multiple labels. See Label expansion for more information.

Note: If the set of labels has high cardinality, this can cause drastic performance issues with Loki. To prevent this from happening, reduce the number of unique label keys and values.

out_of_order_action const: "accept" | const: "rewrite_timestamp" | const: "drop"

Some sources may generate events with timestamps that aren't in chronological order. Even though the sink sorts the events before sending them to Loki, there is a chance that another event could come in that is out of order with the latest events sent to Loki. Prior to Loki 2.4.0, this was not supported and would result in an error during the push request.

If you're using Loki 2.4.0 or newer, Accept is the preferred action, which lets Loki handle any necessary sorting/reordering. If you're using an earlier version, then you must use Drop or RewriteTimestamp depending on which option makes the most sense for your use case.

Default: "accept"
path string

The path to use in the URL of the Loki instance.

Default: "/loki/api/v1/push"
remove_label_fields boolean

Whether or not to delete fields from the event when they are used as labels.

Default: false
remove_structured_metadata_fields boolean

Whether or not to delete fields from the event when they are used in structured metadata.

Default: false
remove_timestamp boolean

The timestamp is still sent as event metadata for Loki to use for indexing.

Default: true
request object

Various settings can be configured, such as concurrency and rate limits, timeouts, retry behavior, etc.

Note that the retry backoff policy follows the Fibonacci sequence.

9 nested properties
adaptive_concurrency object

These parameters typically do not require changes from the default, and incorrect values can lead to meta-stable or unstable performance and sink behavior. Proceed with caution.

5 nested properties
decrease_ratio number

Valid values are greater than 0 and less than 1. Smaller values cause the algorithm to scale back rapidly when latency increases.

Note that the new limit is rounded down after applying this ratio.

Default: 0.9
min=0.0max=1.0
ewma_alpha number

Valid values are greater than 0 and less than 1.

ARC uses an exponentially weighted moving average (EWMA) of past RTT measurements as a reference to compare with the current RTT. Smaller values cause this reference to adjust more slowly, which may be useful if a service has unusually high response variability.

Default: 0.4
min=0.0max=1.0
initial_concurrency integer

It is recommended to set this value to your service's average limit if you're seeing that it takes a long time to ramp up adaptive concurrency after a restart. You can find this value by looking at the adaptive_concurrency_limit metric.

Default: 1
min=1.0max=9007199254740990.0
max_concurrency_limit integer

The adaptive request concurrency limit will not go above this bound. This is put in place as a safeguard.

Default: 200
min=1.0max=9007199254740990.0
rtt_deviation_scale number

Valid values are greater than or equal to 0, and we expect reasonable values to range from 1.0 to 3.0.

When calculating the past RTT average, we also compute a secondary “deviation” value that indicates how variable those values are. We use that deviation when comparing the past RTT average to the current measurements, so we can ignore increases in RTT that are within an expected range. This factor is used to scale up the deviation to an appropriate range. Larger values cause the algorithm to ignore larger increases in the RTT.

Default: 2.5
min=0.0max=9007199254740990.0
concurrency const: "none" | const: "adaptive" | integer

Configuration for outbound request concurrency.

This can be set either to one of the below enum values or to a positive integer, which denotes a fixed concurrency limit.

rate_limit_duration_secs integer

The time window used for the rate_limit_num option.

Default: 1
min=0.0max=9007199254740990.0
rate_limit_num integer

The maximum number of requests allowed within the rate_limit_duration_secs time window.

Default: 9223372036854775807
min=0.0max=9007199254740990.0
retry_attempts integer

The maximum number of retries to make for failed requests.

Default: 9223372036854775807
min=0.0max=9007199254740990.0
retry_initial_backoff_secs integer

After the first retry has failed, the fibonacci sequence is used to select future backoffs.

Default: 1
min=1.0max=9007199254740990.0
retry_jitter_mode const: "None" | const: "Full"

The jitter mode to use for retry backoff behavior.

retry_max_duration_secs integer

The maximum amount of time to wait between retries.

Default: 30
min=1.0max=9007199254740990.0
timeout_secs integer

Datadog highly recommends that you do not lower this value below the service's internal timeout, as this could create orphaned requests, pile on retries, and result in duplicate data downstream.

Default: 60
min=0.0max=9007199254740990.0
structured_metadata Record<string, string>

Both keys and values are templateable, which enables you to attach dynamic structured metadata to events.

Valid metadata keys include *, and prefixes ending with *, to allow for the expansion of objects into multiple metadata entries. This follows the same logic as Label expansion.

Default:
{}

When running Loki locally, a tenant ID is not required.

vector::sinks::mezmo::LogdnaConfig object

Configuration for the mezmo (formerly logdna) sink.

api_key string required

Wrapper for sensitive strings containing credentials

hostname string required

In many cases, components can be configured so that part of the component's functionality can be customized on a per-event basis. For example, you have a sink that writes events to a file and you want to specify which file an event should go to by using an event field as part of the input to the filename used.

By using Template, users can specify either fixed strings or templated strings. Templated strings use a common syntax to refer to fields in an event that is used as the input data when rendering the template. An example of a fixed string is my-file.log. An example of a template string is my-file-{{key}}.log, where {{key}} is the key's value when the template is rendered into a string.

acknowledgements object

See End-to-end Acknowledgements for more information on how event acknowledgement is handled.

1 nested properties
enabled boolean | null

When enabled for a sink, any source connected to that sink, where the source supports end-to-end acknowledgements as well, waits for events to be acknowledged by all connected sinks before acknowledging them at the source.

Enabling or disabling acknowledgements at the sink level takes precedence over any global acknowledgements configuration.

Event batching behavior.

Default:
{
  "max_bytes": null,
  "max_events": null,
  "timeout_secs": null
}
default_app string

The default app that is set for events that do not contain a file or app field.

Default: "vector"
default_env string

The default environment that is set for events that do not contain an env field.

Default: "production"
encoding object

Transformations to prepare an event for serialization.

3 nested properties
except_fields array | null

List of fields that are excluded from the encoded event.

Default: null
only_fields array | null

List of fields that are included in the encoded event.

Default: null
timestamp_format null | const: "unix" | const: "rfc3339" | const: "unix_ms" | const: "unix_us" | const: "unix_ns" | const: "unix_float"

Format used for timestamp fields.

Default: null
endpoint string
ip string | null

The IP address that is attached to each batch of events.

mac string | null

The MAC address that is attached to each batch of events.

request object

Various settings can be configured, such as concurrency and rate limits, timeouts, retry behavior, etc.

Note that the retry backoff policy follows the Fibonacci sequence.

9 nested properties
adaptive_concurrency object

These parameters typically do not require changes from the default, and incorrect values can lead to meta-stable or unstable performance and sink behavior. Proceed with caution.

5 nested properties
decrease_ratio number

Valid values are greater than 0 and less than 1. Smaller values cause the algorithm to scale back rapidly when latency increases.

Note that the new limit is rounded down after applying this ratio.

Default: 0.9
min=0.0max=1.0
ewma_alpha number

Valid values are greater than 0 and less than 1.

ARC uses an exponentially weighted moving average (EWMA) of past RTT measurements as a reference to compare with the current RTT. Smaller values cause this reference to adjust more slowly, which may be useful if a service has unusually high response variability.

Default: 0.4
min=0.0max=1.0
initial_concurrency integer

It is recommended to set this value to your service's average limit if you're seeing that it takes a long time to ramp up adaptive concurrency after a restart. You can find this value by looking at the adaptive_concurrency_limit metric.

Default: 1
min=1.0max=9007199254740990.0
max_concurrency_limit integer

The adaptive request concurrency limit will not go above this bound. This is put in place as a safeguard.

Default: 200
min=1.0max=9007199254740990.0
rtt_deviation_scale number

Valid values are greater than or equal to 0, and we expect reasonable values to range from 1.0 to 3.0.

When calculating the past RTT average, we also compute a secondary “deviation” value that indicates how variable those values are. We use that deviation when comparing the past RTT average to the current measurements, so we can ignore increases in RTT that are within an expected range. This factor is used to scale up the deviation to an appropriate range. Larger values cause the algorithm to ignore larger increases in the RTT.

Default: 2.5
min=0.0max=9007199254740990.0
concurrency const: "none" | const: "adaptive" | integer

Configuration for outbound request concurrency.

This can be set either to one of the below enum values or to a positive integer, which denotes a fixed concurrency limit.

rate_limit_duration_secs integer

The time window used for the rate_limit_num option.

Default: 1
min=0.0max=9007199254740990.0
rate_limit_num integer

The maximum number of requests allowed within the rate_limit_duration_secs time window.

Default: 9223372036854775807
min=0.0max=9007199254740990.0
retry_attempts integer

The maximum number of retries to make for failed requests.

Default: 9223372036854775807
min=0.0max=9007199254740990.0
retry_initial_backoff_secs integer

After the first retry has failed, the fibonacci sequence is used to select future backoffs.

Default: 1
min=1.0max=9007199254740990.0
retry_jitter_mode const: "None" | const: "Full"

The jitter mode to use for retry backoff behavior.

retry_max_duration_secs integer

The maximum amount of time to wait between retries.

Default: 30
min=1.0max=9007199254740990.0
timeout_secs integer

Datadog highly recommends that you do not lower this value below the service's internal timeout, as this could create orphaned requests, pile on retries, and result in duplicate data downstream.

Default: 60
min=0.0max=9007199254740990.0
tags array | null

The tags that are attached to each batch of events.

vector::sinks::mezmo::MezmoConfig object

Configuration for the mezmo (formerly logdna) sink.

api_key string required

Wrapper for sensitive strings containing credentials

hostname string required

In many cases, components can be configured so that part of the component's functionality can be customized on a per-event basis. For example, you have a sink that writes events to a file and you want to specify which file an event should go to by using an event field as part of the input to the filename used.

By using Template, users can specify either fixed strings or templated strings. Templated strings use a common syntax to refer to fields in an event that is used as the input data when rendering the template. An example of a fixed string is my-file.log. An example of a template string is my-file-{{key}}.log, where {{key}} is the key's value when the template is rendered into a string.

acknowledgements object

See End-to-end Acknowledgements for more information on how event acknowledgement is handled.

1 nested properties
enabled boolean | null

When enabled for a sink, any source connected to that sink, where the source supports end-to-end acknowledgements as well, waits for events to be acknowledged by all connected sinks before acknowledging them at the source.

Enabling or disabling acknowledgements at the sink level takes precedence over any global acknowledgements configuration.

Event batching behavior.

Default:
{
  "max_bytes": null,
  "max_events": null,
  "timeout_secs": null
}
default_app string

The default app that is set for events that do not contain a file or app field.

Default: "vector"
default_env string

The default environment that is set for events that do not contain an env field.

Default: "production"
encoding object

Transformations to prepare an event for serialization.

3 nested properties
except_fields array | null

List of fields that are excluded from the encoded event.

Default: null
only_fields array | null

List of fields that are included in the encoded event.

Default: null
timestamp_format null | const: "unix" | const: "rfc3339" | const: "unix_ms" | const: "unix_us" | const: "unix_ns" | const: "unix_float"

Format used for timestamp fields.

Default: null
endpoint string
ip string | null

The IP address that is attached to each batch of events.

mac string | null

The MAC address that is attached to each batch of events.

request object

Various settings can be configured, such as concurrency and rate limits, timeouts, retry behavior, etc.

Note that the retry backoff policy follows the Fibonacci sequence.

9 nested properties
adaptive_concurrency object

These parameters typically do not require changes from the default, and incorrect values can lead to meta-stable or unstable performance and sink behavior. Proceed with caution.

5 nested properties
decrease_ratio number

Valid values are greater than 0 and less than 1. Smaller values cause the algorithm to scale back rapidly when latency increases.

Note that the new limit is rounded down after applying this ratio.

Default: 0.9
min=0.0max=1.0
ewma_alpha number

Valid values are greater than 0 and less than 1.

ARC uses an exponentially weighted moving average (EWMA) of past RTT measurements as a reference to compare with the current RTT. Smaller values cause this reference to adjust more slowly, which may be useful if a service has unusually high response variability.

Default: 0.4
min=0.0max=1.0
initial_concurrency integer

It is recommended to set this value to your service's average limit if you're seeing that it takes a long time to ramp up adaptive concurrency after a restart. You can find this value by looking at the adaptive_concurrency_limit metric.

Default: 1
min=1.0max=9007199254740990.0
max_concurrency_limit integer

The adaptive request concurrency limit will not go above this bound. This is put in place as a safeguard.

Default: 200
min=1.0max=9007199254740990.0
rtt_deviation_scale number

Valid values are greater than or equal to 0, and we expect reasonable values to range from 1.0 to 3.0.

When calculating the past RTT average, we also compute a secondary “deviation” value that indicates how variable those values are. We use that deviation when comparing the past RTT average to the current measurements, so we can ignore increases in RTT that are within an expected range. This factor is used to scale up the deviation to an appropriate range. Larger values cause the algorithm to ignore larger increases in the RTT.

Default: 2.5
min=0.0max=9007199254740990.0
concurrency const: "none" | const: "adaptive" | integer

Configuration for outbound request concurrency.

This can be set either to one of the below enum values or to a positive integer, which denotes a fixed concurrency limit.

rate_limit_duration_secs integer

The time window used for the rate_limit_num option.

Default: 1
min=0.0max=9007199254740990.0
rate_limit_num integer

The maximum number of requests allowed within the rate_limit_duration_secs time window.

Default: 9223372036854775807
min=0.0max=9007199254740990.0
retry_attempts integer

The maximum number of retries to make for failed requests.

Default: 9223372036854775807
min=0.0max=9007199254740990.0
retry_initial_backoff_secs integer

After the first retry has failed, the fibonacci sequence is used to select future backoffs.

Default: 1
min=1.0max=9007199254740990.0
retry_jitter_mode const: "None" | const: "Full"

The jitter mode to use for retry backoff behavior.

retry_max_duration_secs integer

The maximum amount of time to wait between retries.

Default: 30
min=1.0max=9007199254740990.0
timeout_secs integer

Datadog highly recommends that you do not lower this value below the service's internal timeout, as this could create orphaned requests, pile on retries, and result in duplicate data downstream.

Default: 60
min=0.0max=9007199254740990.0
tags array | null

The tags that are attached to each batch of events.

vector::sinks::mqtt::config::MqttSinkConfig object

Configuration for the mqtt sink

encoding required

Configures how events are encoded into raw bytes.

All of: object | object | object | object | object | object object | object | object | object | object | object, object object
host string required

MQTT server address (The broker’s domain name or IP address).

topic string required

In many cases, components can be configured so that part of the component's functionality can be customized on a per-event basis. For example, you have a sink that writes events to a file and you want to specify which file an event should go to by using an event field as part of the input to the filename used.

By using Template, users can specify either fixed strings or templated strings. Templated strings use a common syntax to refer to fields in an event that is used as the input data when rendering the template. An example of a fixed string is my-file.log. An example of a template string is my-file-{{key}}.log, where {{key}} is the key's value when the template is rendered into a string.

acknowledgements object

See End-to-end Acknowledgements for more information on how event acknowledgement is handled.

1 nested properties
enabled boolean | null

When enabled for a sink, any source connected to that sink, where the source supports end-to-end acknowledgements as well, waits for events to be acknowledged by all connected sinks before acknowledging them at the source.

Enabling or disabling acknowledgements at the sink level takes precedence over any global acknowledgements configuration.

clean_session boolean

If set to true, the MQTT session is cleaned on login.

Default: false
client_id string | null

MQTT client ID.

keep_alive integer

Connection keep-alive interval.

Default: 60
min=0.0max=65535.0
password string | null

MQTT password.

port integer

TCP port of the MQTT server to connect to.

Default: 1883
min=0.0max=65535.0
quality_of_service const: "atleastonce" | const: "atmostonce" | const: "exactlyonce"

Supported Quality of Service types for MQTT.

Default: "atleastonce"
retain boolean

Whether the messages should be retained by the server

Default: false
user string | null

MQTT username.

vector::sinks::nats::config::NatsSinkConfig object

Configuration for the nats sink.

encoding required

Configures how events are encoded into raw bytes.

All of: object | object | object | object | object | object object | object | object | object | object | object, object object
subject string required

In many cases, components can be configured so that part of the component's functionality can be customized on a per-event basis. For example, you have a sink that writes events to a file and you want to specify which file an event should go to by using an event field as part of the input to the filename used.

By using Template, users can specify either fixed strings or templated strings. Templated strings use a common syntax to refer to fields in an event that is used as the input data when rendering the template. An example of a fixed string is my-file.log. An example of a template string is my-file-{{key}}.log, where {{key}} is the key's value when the template is rendered into a string.

url string required

The URL must take the form of nats://server:port. If the port is not specified it defaults to 4222.

acknowledgements object

See End-to-end Acknowledgements for more information on how event acknowledgement is handled.

1 nested properties
enabled boolean | null

When enabled for a sink, any source connected to that sink, where the source supports end-to-end acknowledgements as well, waits for events to be acknowledged by all connected sinks before acknowledging them at the source.

Enabling or disabling acknowledgements at the sink level takes precedence over any global acknowledgements configuration.

connection_name string
Default: "vector"
jetstream boolean

If set, the subject must belong to an existing JetStream stream.

Default: false
request object

Various settings can be configured, such as concurrency and rate limits, timeouts, retry behavior, etc.

Note that the retry backoff policy follows the Fibonacci sequence.

Default:
{
  "timeout_secs": 60,
  "rate_limit_duration_secs": 1,
  "rate_limit_num": 9223372036854775807,
  "retry_attempts": 9223372036854775807,
  "retry_max_duration_secs": 30,
  "retry_initial_backoff_secs": 1,
  "retry_jitter_mode": "Full",
  "adaptive_concurrency": {
    "initial_concurrency": 1,
    "decrease_ratio": 0.9,
    "ewma_alpha": 0.4,
    "rtt_deviation_scale": 2.5,
    "max_concurrency_limit": 200
  }
}
9 nested properties
adaptive_concurrency object

These parameters typically do not require changes from the default, and incorrect values can lead to meta-stable or unstable performance and sink behavior. Proceed with caution.

5 nested properties
decrease_ratio number

Valid values are greater than 0 and less than 1. Smaller values cause the algorithm to scale back rapidly when latency increases.

Note that the new limit is rounded down after applying this ratio.

Default: 0.9
min=0.0max=1.0
ewma_alpha number

Valid values are greater than 0 and less than 1.

ARC uses an exponentially weighted moving average (EWMA) of past RTT measurements as a reference to compare with the current RTT. Smaller values cause this reference to adjust more slowly, which may be useful if a service has unusually high response variability.

Default: 0.4
min=0.0max=1.0
initial_concurrency integer

It is recommended to set this value to your service's average limit if you're seeing that it takes a long time to ramp up adaptive concurrency after a restart. You can find this value by looking at the adaptive_concurrency_limit metric.

Default: 1
min=1.0max=9007199254740990.0
max_concurrency_limit integer

The adaptive request concurrency limit will not go above this bound. This is put in place as a safeguard.

Default: 200
min=1.0max=9007199254740990.0
rtt_deviation_scale number

Valid values are greater than or equal to 0, and we expect reasonable values to range from 1.0 to 3.0.

When calculating the past RTT average, we also compute a secondary “deviation” value that indicates how variable those values are. We use that deviation when comparing the past RTT average to the current measurements, so we can ignore increases in RTT that are within an expected range. This factor is used to scale up the deviation to an appropriate range. Larger values cause the algorithm to ignore larger increases in the RTT.

Default: 2.5
min=0.0max=9007199254740990.0
concurrency const: "none" | const: "adaptive" | integer

Configuration for outbound request concurrency.

This can be set either to one of the below enum values or to a positive integer, which denotes a fixed concurrency limit.

rate_limit_duration_secs integer

The time window used for the rate_limit_num option.

Default: 1
min=0.0max=9007199254740990.0
rate_limit_num integer

The maximum number of requests allowed within the rate_limit_duration_secs time window.

Default: 9223372036854775807
min=0.0max=9007199254740990.0
retry_attempts integer

The maximum number of retries to make for failed requests.

Default: 9223372036854775807
min=0.0max=9007199254740990.0
retry_initial_backoff_secs integer

After the first retry has failed, the fibonacci sequence is used to select future backoffs.

Default: 1
min=1.0max=9007199254740990.0
retry_jitter_mode const: "None" | const: "Full"

The jitter mode to use for retry backoff behavior.

retry_max_duration_secs integer

The maximum amount of time to wait between retries.

Default: 30
min=1.0max=9007199254740990.0
timeout_secs integer

Datadog highly recommends that you do not lower this value below the service's internal timeout, as this could create orphaned requests, pile on retries, and result in duplicate data downstream.

Default: 60
min=0.0max=9007199254740990.0
vector::sinks::new_relic::config::NewRelicConfig object

Configuration for the new_relic sink.

account_id string required

Wrapper for sensitive strings containing credentials

api const: "events" | const: "metrics" | const: "logs" required

New Relic API endpoint.

license_key string required

Wrapper for sensitive strings containing credentials

acknowledgements object

See End-to-end Acknowledgements for more information on how event acknowledgement is handled.

1 nested properties
enabled boolean | null

When enabled for a sink, any source connected to that sink, where the source supports end-to-end acknowledgements as well, waits for events to be acknowledged by all connected sinks before acknowledging them at the source.

Enabling or disabling acknowledgements at the sink level takes precedence over any global acknowledgements configuration.

batch object

Event batching behavior.

Default:
{
  "max_bytes": null,
  "max_events": null,
  "timeout_secs": null
}
3 nested properties
max_bytes integer | null

This is based on the uncompressed size of the batched events, before they are serialized/compressed.

Default: 1000000
min=0.0max=9007199254740990.0
max_events integer | null

The maximum size of a batch before it is flushed.

Default: 100
min=0.0max=9007199254740990.0
timeout_secs number | null

The maximum age of a batch before it is flushed.

Default: 1.0
min=-9007199254740990.0max=9007199254740990.0
compression const: "none" | const: "gzip" | const: "zlib" | const: "zstd" | const: "snappy" | object

All compression algorithms use the default compression level unless otherwise specified.

encoding object

Transformations to prepare an event for serialization.

3 nested properties
except_fields array | null

List of fields that are excluded from the encoded event.

Default: null
only_fields array | null

List of fields that are included in the encoded event.

Default: null
timestamp_format null | const: "unix" | const: "rfc3339" | const: "unix_ms" | const: "unix_us" | const: "unix_ns" | const: "unix_float"

Format used for timestamp fields.

Default: null
region null | const: "us" | const: "eu"

New Relic region.

request object

Various settings can be configured, such as concurrency and rate limits, timeouts, retry behavior, etc.

Note that the retry backoff policy follows the Fibonacci sequence.

9 nested properties
adaptive_concurrency object

These parameters typically do not require changes from the default, and incorrect values can lead to meta-stable or unstable performance and sink behavior. Proceed with caution.

5 nested properties
decrease_ratio number

Valid values are greater than 0 and less than 1. Smaller values cause the algorithm to scale back rapidly when latency increases.

Note that the new limit is rounded down after applying this ratio.

Default: 0.9
min=0.0max=1.0
ewma_alpha number

Valid values are greater than 0 and less than 1.

ARC uses an exponentially weighted moving average (EWMA) of past RTT measurements as a reference to compare with the current RTT. Smaller values cause this reference to adjust more slowly, which may be useful if a service has unusually high response variability.

Default: 0.4
min=0.0max=1.0
initial_concurrency integer

It is recommended to set this value to your service's average limit if you're seeing that it takes a long time to ramp up adaptive concurrency after a restart. You can find this value by looking at the adaptive_concurrency_limit metric.

Default: 1
min=1.0max=9007199254740990.0
max_concurrency_limit integer

The adaptive request concurrency limit will not go above this bound. This is put in place as a safeguard.

Default: 200
min=1.0max=9007199254740990.0
rtt_deviation_scale number

Valid values are greater than or equal to 0, and we expect reasonable values to range from 1.0 to 3.0.

When calculating the past RTT average, we also compute a secondary “deviation” value that indicates how variable those values are. We use that deviation when comparing the past RTT average to the current measurements, so we can ignore increases in RTT that are within an expected range. This factor is used to scale up the deviation to an appropriate range. Larger values cause the algorithm to ignore larger increases in the RTT.

Default: 2.5
min=0.0max=9007199254740990.0
concurrency const: "none" | const: "adaptive" | integer

Configuration for outbound request concurrency.

This can be set either to one of the below enum values or to a positive integer, which denotes a fixed concurrency limit.

rate_limit_duration_secs integer

The time window used for the rate_limit_num option.

Default: 1
min=0.0max=9007199254740990.0
rate_limit_num integer

The maximum number of requests allowed within the rate_limit_duration_secs time window.

Default: 9223372036854775807
min=0.0max=9007199254740990.0
retry_attempts integer

The maximum number of retries to make for failed requests.

Default: 9223372036854775807
min=0.0max=9007199254740990.0
retry_initial_backoff_secs integer

After the first retry has failed, the fibonacci sequence is used to select future backoffs.

Default: 1
min=1.0max=9007199254740990.0
retry_jitter_mode const: "None" | const: "Full"

The jitter mode to use for retry backoff behavior.

retry_max_duration_secs integer

The maximum amount of time to wait between retries.

Default: 30
min=1.0max=9007199254740990.0
timeout_secs integer

Datadog highly recommends that you do not lower this value below the service's internal timeout, as this could create orphaned requests, pile on retries, and result in duplicate data downstream.

Default: 60
min=0.0max=9007199254740990.0
vector::sinks::opentelemetry::OpenTelemetryConfig object

Configuration for the OpenTelemetry sink.

protocol required

Protocol configuration

vector::sinks::papertrail::PapertrailConfig object

Configuration for the papertrail sink.

encoding required

Configures how events are encoded into raw bytes.

All of: object | object | object | object | object | object object | object | object | object | object | object, object object
endpoint string required
acknowledgements object

See End-to-end Acknowledgements for more information on how event acknowledgement is handled.

1 nested properties
enabled boolean | null

When enabled for a sink, any source connected to that sink, where the source supports end-to-end acknowledgements as well, waits for events to be acknowledged by all connected sinks before acknowledging them at the source.

Enabling or disabling acknowledgements at the sink level takes precedence over any global acknowledgements configuration.

process string

In many cases, components can be configured so that part of the component's functionality can be customized on a per-event basis. For example, you have a sink that writes events to a file and you want to specify which file an event should go to by using an event field as part of the input to the filename used.

By using Template, users can specify either fixed strings or templated strings. Templated strings use a common syntax to refer to fields in an event that is used as the input data when rendering the template. An example of a fixed string is my-file.log. An example of a template string is my-file-{{key}}.log, where {{key}} is the key's value when the template is rendered into a string.

send_buffer_bytes integer | null

Configures the send buffer size using the SO_SNDBUF option on the socket.

min=0.0max=9007199254740990.0
vector::sinks::prometheus::exporter::PrometheusExporterConfig object

Configuration for the prometheus_exporter sink.

acknowledgements object

See End-to-end Acknowledgements for more information on how event acknowledgement is handled.

1 nested properties
enabled boolean | null

When enabled for a sink, any source connected to that sink, where the source supports end-to-end acknowledgements as well, waits for events to be acknowledged by all connected sinks before acknowledging them at the source.

Enabling or disabling acknowledgements at the sink level takes precedence over any global acknowledgements configuration.

address string

An internet socket address, either IPv4 or IPv6.

buckets number[]
Default:
[
  0.005,
  0.01,
  0.025,
  0.05,
  0.1,
  0.25,
  0.5,
  1.0,
  2.5,
  5.0,
  10.0
]
default_namespace string | null

This namespace is only used if a metric has no existing namespace. When a namespace is present, it is used as a prefix to the metric name, and separated with an underscore (_).

It should follow the Prometheus naming conventions.

distributions_as_summaries boolean

While distributions as a lossless way to represent a set of samples for a metric is supported, Prometheus clients (the application being scraped, which is this sink) must aggregate locally into either an aggregated histogram or aggregated summary.

Default: false
flush_period_secs integer

A span of time, in whole seconds.

min=0.0max=9007199254740990.0
quantiles number[]
Default:
[
  0.5,
  0.75,
  0.9,
  0.95,
  0.99
]
suppress_timestamp boolean

This can sometimes be useful when the source of metrics leads to their timestamps being too far in the past for Prometheus to allow them, such as when aggregating metrics over long time periods, or when replaying old metrics from a disk buffer.

Default: false
vector::sinks::prometheus::remote_write::config::RemoteWriteConfig object

Configuration for the prometheus_remote_write sink.

endpoint string required

The endpoint should include the scheme and the path to write to.

acknowledgements object

See End-to-end Acknowledgements for more information on how event acknowledgement is handled.

1 nested properties
enabled boolean | null

When enabled for a sink, any source connected to that sink, where the source supports end-to-end acknowledgements as well, waits for events to be acknowledged by all connected sinks before acknowledging them at the source.

Enabling or disabling acknowledgements at the sink level takes precedence over any global acknowledgements configuration.

auth null | object | object

Authentication strategies.

batch

The batch config for remote write.

Default:
{
  "max_bytes": null,
  "max_events": null,
  "timeout_secs": null,
  "aggregate": true
}
All of: object object, object object
buckets number[]
Default:
[
  0.005,
  0.01,
  0.025,
  0.05,
  0.1,
  0.25,
  0.5,
  1.0,
  2.5,
  5.0,
  10.0
]
compression const: "none" | const: "gzip" | const: "zlib" | const: "zstd" | const: "snappy" | object

All compression algorithms use the default compression level unless otherwise specified.

default_namespace string | null

This namespace is only used if a metric has no existing namespace. When a namespace is present, it is used as a prefix to the metric name, and separated with an underscore (_).

It should follow the Prometheus naming conventions.

quantiles number[]
Default:
[
  0.5,
  0.75,
  0.9,
  0.95,
  0.99
]
request object

Various settings can be configured, such as concurrency and rate limits, timeouts, retry behavior, etc.

Note that the retry backoff policy follows the Fibonacci sequence.

9 nested properties
adaptive_concurrency object

These parameters typically do not require changes from the default, and incorrect values can lead to meta-stable or unstable performance and sink behavior. Proceed with caution.

5 nested properties
decrease_ratio number

Valid values are greater than 0 and less than 1. Smaller values cause the algorithm to scale back rapidly when latency increases.

Note that the new limit is rounded down after applying this ratio.

Default: 0.9
min=0.0max=1.0
ewma_alpha number

Valid values are greater than 0 and less than 1.

ARC uses an exponentially weighted moving average (EWMA) of past RTT measurements as a reference to compare with the current RTT. Smaller values cause this reference to adjust more slowly, which may be useful if a service has unusually high response variability.

Default: 0.4
min=0.0max=1.0
initial_concurrency integer

It is recommended to set this value to your service's average limit if you're seeing that it takes a long time to ramp up adaptive concurrency after a restart. You can find this value by looking at the adaptive_concurrency_limit metric.

Default: 1
min=1.0max=9007199254740990.0
max_concurrency_limit integer

The adaptive request concurrency limit will not go above this bound. This is put in place as a safeguard.

Default: 200
min=1.0max=9007199254740990.0
rtt_deviation_scale number

Valid values are greater than or equal to 0, and we expect reasonable values to range from 1.0 to 3.0.

When calculating the past RTT average, we also compute a secondary “deviation” value that indicates how variable those values are. We use that deviation when comparing the past RTT average to the current measurements, so we can ignore increases in RTT that are within an expected range. This factor is used to scale up the deviation to an appropriate range. Larger values cause the algorithm to ignore larger increases in the RTT.

Default: 2.5
min=0.0max=9007199254740990.0
concurrency const: "none" | const: "adaptive" | integer

Configuration for outbound request concurrency.

This can be set either to one of the below enum values or to a positive integer, which denotes a fixed concurrency limit.

rate_limit_duration_secs integer

The time window used for the rate_limit_num option.

Default: 1
min=0.0max=9007199254740990.0
rate_limit_num integer

The maximum number of requests allowed within the rate_limit_duration_secs time window.

Default: 9223372036854775807
min=0.0max=9007199254740990.0
retry_attempts integer

The maximum number of retries to make for failed requests.

Default: 9223372036854775807
min=0.0max=9007199254740990.0
retry_initial_backoff_secs integer

After the first retry has failed, the fibonacci sequence is used to select future backoffs.

Default: 1
min=1.0max=9007199254740990.0
retry_jitter_mode const: "None" | const: "Full"

The jitter mode to use for retry backoff behavior.

retry_max_duration_secs integer

The maximum amount of time to wait between retries.

Default: 30
min=1.0max=9007199254740990.0
timeout_secs integer

Datadog highly recommends that you do not lower this value below the service's internal timeout, as this could create orphaned requests, pile on retries, and result in duplicate data downstream.

Default: 60
min=0.0max=9007199254740990.0

If set, a header named X-Scope-OrgID is added to outgoing requests with the value of this setting.

This may be used by Cortex or other remote services to identify the tenant making the request.

Default: null
vector::sinks::pulsar::config::PulsarSinkConfig object

Configuration for the pulsar sink.

encoding required

Configures how events are encoded into raw bytes.

All of: object | object | object | object | object | object object | object | object | object | object | object, object object
endpoint string required

The endpoint should specify the pulsar protocol and port.

topic string required

In many cases, components can be configured so that part of the component's functionality can be customized on a per-event basis. For example, you have a sink that writes events to a file and you want to specify which file an event should go to by using an event field as part of the input to the filename used.

By using Template, users can specify either fixed strings or templated strings. Templated strings use a common syntax to refer to fields in an event that is used as the input data when rendering the template. An example of a fixed string is my-file.log. An example of a template string is my-file-{{key}}.log, where {{key}} is the key's value when the template is rendered into a string.

acknowledgements object

See End-to-end Acknowledgements for more information on how event acknowledgement is handled.

1 nested properties
enabled boolean | null

When enabled for a sink, any source connected to that sink, where the source supports end-to-end acknowledgements as well, waits for events to be acknowledged by all connected sinks before acknowledging them at the source.

Enabling or disabling acknowledgements at the sink level takes precedence over any global acknowledgements configuration.

auth null | object

Authentication configuration.

batch object

Event batching behavior.

Default:
{
  "max_events": null,
  "max_bytes": null
}
2 nested properties
max_bytes integer | null

The maximum size of a batch before it is flushed.

min=0.0max=9007199254740990.0
max_events integer | null

Note this is an unsigned 32 bit integer which is a smaller capacity than many of the other sink batch settings.

min=0.0max=4294967295.0
compression const: "none" | const: "lz4" | const: "zlib" | const: "zstd" | const: "snappy"

Supported compression types for Pulsar.

Default: "none"
connection_retry_options null | object

Custom connection retry options configuration for the Pulsar client.

Default: null

If the field does not exist in the log event or metric tags, a blank value will be used.

If omitted, the key is not sent.

Pulsar uses a hash of the key to choose the topic-partition or uses round-robin if the record has no key.

producer_name string | null

The name of the producer. If not specified, the default name assigned by Pulsar is used.

If omitted, no properties will be written.

vector::sinks::redis::config::RedisSinkConfig object

Configuration for the redis sink.

encoding required

Configures how events are encoded into raw bytes.

All of: object | object | object | object | object | object object | object | object | object | object | object, object object
endpoint string required

The URL must take the form of protocol://server:port/db where the protocol can either be redis or rediss for connections secured via TLS.

key string required

In many cases, components can be configured so that part of the component's functionality can be customized on a per-event basis. For example, you have a sink that writes events to a file and you want to specify which file an event should go to by using an event field as part of the input to the filename used.

By using Template, users can specify either fixed strings or templated strings. Templated strings use a common syntax to refer to fields in an event that is used as the input data when rendering the template. An example of a fixed string is my-file.log. An example of a template string is my-file-{{key}}.log, where {{key}} is the key's value when the template is rendered into a string.

acknowledgements object

See End-to-end Acknowledgements for more information on how event acknowledgement is handled.

1 nested properties
enabled boolean | null

When enabled for a sink, any source connected to that sink, where the source supports end-to-end acknowledgements as well, waits for events to be acknowledged by all connected sinks before acknowledging them at the source.

Enabling or disabling acknowledgements at the sink level takes precedence over any global acknowledgements configuration.

batch object

Event batching behavior.

Default:
{
  "max_bytes": null,
  "max_events": null,
  "timeout_secs": null
}
3 nested properties
max_bytes integer | null

This is based on the uncompressed size of the batched events, before they are serialized/compressed.

Default: null
min=0.0max=9007199254740990.0
max_events integer | null

The maximum size of a batch before it is flushed.

Default: 1
min=0.0max=9007199254740990.0
timeout_secs number | null

The maximum age of a batch before it is flushed.

Default: 1.0
min=-9007199254740990.0max=9007199254740990.0
data_type const: "list" | const: "channel"

Redis data type to store messages in.

Default: "list"
list_option null | object

List-specific options.

request object

Various settings can be configured, such as concurrency and rate limits, timeouts, retry behavior, etc.

Note that the retry backoff policy follows the Fibonacci sequence.

Default:
{
  "timeout_secs": 60,
  "rate_limit_duration_secs": 1,
  "rate_limit_num": 9223372036854775807,
  "retry_attempts": 9223372036854775807,
  "retry_max_duration_secs": 30,
  "retry_initial_backoff_secs": 1,
  "retry_jitter_mode": "Full",
  "adaptive_concurrency": {
    "initial_concurrency": 1,
    "decrease_ratio": 0.9,
    "ewma_alpha": 0.4,
    "rtt_deviation_scale": 2.5,
    "max_concurrency_limit": 200
  }
}
9 nested properties
adaptive_concurrency object

These parameters typically do not require changes from the default, and incorrect values can lead to meta-stable or unstable performance and sink behavior. Proceed with caution.

5 nested properties
decrease_ratio number

Valid values are greater than 0 and less than 1. Smaller values cause the algorithm to scale back rapidly when latency increases.

Note that the new limit is rounded down after applying this ratio.

Default: 0.9
min=0.0max=1.0
ewma_alpha number

Valid values are greater than 0 and less than 1.

ARC uses an exponentially weighted moving average (EWMA) of past RTT measurements as a reference to compare with the current RTT. Smaller values cause this reference to adjust more slowly, which may be useful if a service has unusually high response variability.

Default: 0.4
min=0.0max=1.0
initial_concurrency integer

It is recommended to set this value to your service's average limit if you're seeing that it takes a long time to ramp up adaptive concurrency after a restart. You can find this value by looking at the adaptive_concurrency_limit metric.

Default: 1
min=1.0max=9007199254740990.0
max_concurrency_limit integer

The adaptive request concurrency limit will not go above this bound. This is put in place as a safeguard.

Default: 200
min=1.0max=9007199254740990.0
rtt_deviation_scale number

Valid values are greater than or equal to 0, and we expect reasonable values to range from 1.0 to 3.0.

When calculating the past RTT average, we also compute a secondary “deviation” value that indicates how variable those values are. We use that deviation when comparing the past RTT average to the current measurements, so we can ignore increases in RTT that are within an expected range. This factor is used to scale up the deviation to an appropriate range. Larger values cause the algorithm to ignore larger increases in the RTT.

Default: 2.5
min=0.0max=9007199254740990.0
concurrency const: "none" | const: "adaptive" | integer

Configuration for outbound request concurrency.

This can be set either to one of the below enum values or to a positive integer, which denotes a fixed concurrency limit.

rate_limit_duration_secs integer

The time window used for the rate_limit_num option.

Default: 1
min=0.0max=9007199254740990.0
rate_limit_num integer

The maximum number of requests allowed within the rate_limit_duration_secs time window.

Default: 9223372036854775807
min=0.0max=9007199254740990.0
retry_attempts integer

The maximum number of retries to make for failed requests.

Default: 9223372036854775807
min=0.0max=9007199254740990.0
retry_initial_backoff_secs integer

After the first retry has failed, the fibonacci sequence is used to select future backoffs.

Default: 1
min=1.0max=9007199254740990.0
retry_jitter_mode const: "None" | const: "Full"

The jitter mode to use for retry backoff behavior.

retry_max_duration_secs integer

The maximum amount of time to wait between retries.

Default: 30
min=1.0max=9007199254740990.0
timeout_secs integer

Datadog highly recommends that you do not lower this value below the service's internal timeout, as this could create orphaned requests, pile on retries, and result in duplicate data downstream.

Default: 60
min=0.0max=9007199254740990.0
vector::sinks::sematext::Region const: "us" | const: "eu"

The Sematext region to send data to.

vector::sinks::sematext::logs::SematextLogsConfig object

Configuration for the sematext_logs sink.

token string required

Wrapper for sensitive strings containing credentials

acknowledgements object

See End-to-end Acknowledgements for more information on how event acknowledgement is handled.

1 nested properties
enabled boolean | null

When enabled for a sink, any source connected to that sink, where the source supports end-to-end acknowledgements as well, waits for events to be acknowledged by all connected sinks before acknowledging them at the source.

Enabling or disabling acknowledgements at the sink level takes precedence over any global acknowledgements configuration.

Event batching behavior.

Default:
{
  "max_bytes": null,
  "max_events": null,
  "timeout_secs": null
}
encoding object

Transformations to prepare an event for serialization.

3 nested properties
except_fields array | null

List of fields that are excluded from the encoded event.

Default: null
only_fields array | null

List of fields that are included in the encoded event.

Default: null
timestamp_format null | const: "unix" | const: "rfc3339" | const: "unix_ms" | const: "unix_us" | const: "unix_ns" | const: "unix_float"

Format used for timestamp fields.

Default: null
endpoint string | null

Setting this option overrides the region option.

region const: "us" | const: "eu"

The Sematext region to send data to.

request object

Various settings can be configured, such as concurrency and rate limits, timeouts, retry behavior, etc.

Note that the retry backoff policy follows the Fibonacci sequence.

9 nested properties
adaptive_concurrency object

These parameters typically do not require changes from the default, and incorrect values can lead to meta-stable or unstable performance and sink behavior. Proceed with caution.

5 nested properties
decrease_ratio number

Valid values are greater than 0 and less than 1. Smaller values cause the algorithm to scale back rapidly when latency increases.

Note that the new limit is rounded down after applying this ratio.

Default: 0.9
min=0.0max=1.0
ewma_alpha number

Valid values are greater than 0 and less than 1.

ARC uses an exponentially weighted moving average (EWMA) of past RTT measurements as a reference to compare with the current RTT. Smaller values cause this reference to adjust more slowly, which may be useful if a service has unusually high response variability.

Default: 0.4
min=0.0max=1.0
initial_concurrency integer

It is recommended to set this value to your service's average limit if you're seeing that it takes a long time to ramp up adaptive concurrency after a restart. You can find this value by looking at the adaptive_concurrency_limit metric.

Default: 1
min=1.0max=9007199254740990.0
max_concurrency_limit integer

The adaptive request concurrency limit will not go above this bound. This is put in place as a safeguard.

Default: 200
min=1.0max=9007199254740990.0
rtt_deviation_scale number

Valid values are greater than or equal to 0, and we expect reasonable values to range from 1.0 to 3.0.

When calculating the past RTT average, we also compute a secondary “deviation” value that indicates how variable those values are. We use that deviation when comparing the past RTT average to the current measurements, so we can ignore increases in RTT that are within an expected range. This factor is used to scale up the deviation to an appropriate range. Larger values cause the algorithm to ignore larger increases in the RTT.

Default: 2.5
min=0.0max=9007199254740990.0
concurrency const: "none" | const: "adaptive" | integer

Configuration for outbound request concurrency.

This can be set either to one of the below enum values or to a positive integer, which denotes a fixed concurrency limit.

rate_limit_duration_secs integer

The time window used for the rate_limit_num option.

Default: 1
min=0.0max=9007199254740990.0
rate_limit_num integer

The maximum number of requests allowed within the rate_limit_duration_secs time window.

Default: 9223372036854775807
min=0.0max=9007199254740990.0
retry_attempts integer

The maximum number of retries to make for failed requests.

Default: 9223372036854775807
min=0.0max=9007199254740990.0
retry_initial_backoff_secs integer

After the first retry has failed, the fibonacci sequence is used to select future backoffs.

Default: 1
min=1.0max=9007199254740990.0
retry_jitter_mode const: "None" | const: "Full"

The jitter mode to use for retry backoff behavior.

retry_max_duration_secs integer

The maximum amount of time to wait between retries.

Default: 30
min=1.0max=9007199254740990.0
timeout_secs integer

Datadog highly recommends that you do not lower this value below the service's internal timeout, as this could create orphaned requests, pile on retries, and result in duplicate data downstream.

Default: 60
min=0.0max=9007199254740990.0
vector::sinks::sematext::metrics::SematextMetricsConfig object

Configuration for the sematext_metrics sink.

default_namespace string required

This namespace is only used if a metric has no existing namespace. When a namespace is present, it is used as a prefix to the metric name, and separated with a period (.).

token string required

Wrapper for sensitive strings containing credentials

acknowledgements object

See End-to-end Acknowledgements for more information on how event acknowledgement is handled.

1 nested properties
enabled boolean | null

When enabled for a sink, any source connected to that sink, where the source supports end-to-end acknowledgements as well, waits for events to be acknowledged by all connected sinks before acknowledging them at the source.

Enabling or disabling acknowledgements at the sink level takes precedence over any global acknowledgements configuration.

batch object

Event batching behavior.

Default:
{
  "max_bytes": null,
  "max_events": null,
  "timeout_secs": null
}
3 nested properties
max_bytes integer | null

This is based on the uncompressed size of the batched events, before they are serialized/compressed.

Default: null
min=0.0max=9007199254740990.0
max_events integer | null

The maximum size of a batch before it is flushed.

Default: 20
min=0.0max=9007199254740990.0
timeout_secs number | null

The maximum age of a batch before it is flushed.

Default: 1.0
min=-9007199254740990.0max=9007199254740990.0
endpoint string | null

Setting this option overrides the region option.

region const: "us" | const: "eu"

The Sematext region to send data to.

request object

Various settings can be configured, such as concurrency and rate limits, timeouts, retry behavior, etc.

Note that the retry backoff policy follows the Fibonacci sequence.

9 nested properties
adaptive_concurrency object

These parameters typically do not require changes from the default, and incorrect values can lead to meta-stable or unstable performance and sink behavior. Proceed with caution.

5 nested properties
decrease_ratio number

Valid values are greater than 0 and less than 1. Smaller values cause the algorithm to scale back rapidly when latency increases.

Note that the new limit is rounded down after applying this ratio.

Default: 0.9
min=0.0max=1.0
ewma_alpha number

Valid values are greater than 0 and less than 1.

ARC uses an exponentially weighted moving average (EWMA) of past RTT measurements as a reference to compare with the current RTT. Smaller values cause this reference to adjust more slowly, which may be useful if a service has unusually high response variability.

Default: 0.4
min=0.0max=1.0
initial_concurrency integer

It is recommended to set this value to your service's average limit if you're seeing that it takes a long time to ramp up adaptive concurrency after a restart. You can find this value by looking at the adaptive_concurrency_limit metric.

Default: 1
min=1.0max=9007199254740990.0
max_concurrency_limit integer

The adaptive request concurrency limit will not go above this bound. This is put in place as a safeguard.

Default: 200
min=1.0max=9007199254740990.0
rtt_deviation_scale number

Valid values are greater than or equal to 0, and we expect reasonable values to range from 1.0 to 3.0.

When calculating the past RTT average, we also compute a secondary “deviation” value that indicates how variable those values are. We use that deviation when comparing the past RTT average to the current measurements, so we can ignore increases in RTT that are within an expected range. This factor is used to scale up the deviation to an appropriate range. Larger values cause the algorithm to ignore larger increases in the RTT.

Default: 2.5
min=0.0max=9007199254740990.0
concurrency const: "none" | const: "adaptive" | integer

Configuration for outbound request concurrency.

This can be set either to one of the below enum values or to a positive integer, which denotes a fixed concurrency limit.

rate_limit_duration_secs integer

The time window used for the rate_limit_num option.

Default: 1
min=0.0max=9007199254740990.0
rate_limit_num integer

The maximum number of requests allowed within the rate_limit_duration_secs time window.

Default: 9223372036854775807
min=0.0max=9007199254740990.0
retry_attempts integer

The maximum number of retries to make for failed requests.

Default: 9223372036854775807
min=0.0max=9007199254740990.0
retry_initial_backoff_secs integer

After the first retry has failed, the fibonacci sequence is used to select future backoffs.

Default: 1
min=1.0max=9007199254740990.0
retry_jitter_mode const: "None" | const: "Full"

The jitter mode to use for retry backoff behavior.

retry_max_duration_secs integer

The maximum amount of time to wait between retries.

Default: 30
min=1.0max=9007199254740990.0
timeout_secs integer

Datadog highly recommends that you do not lower this value below the service's internal timeout, as this could create orphaned requests, pile on retries, and result in duplicate data downstream.

Default: 60
min=0.0max=9007199254740990.0
vector::sinks::socket::SocketSinkConfig

Configuration for the socket sink.

vector::sinks::splunk_hec::common::acknowledgements::HecClientAcknowledgementsConfig

Splunk HEC acknowledgement configuration.

vector::sinks::splunk_hec::logs::config::HecLogsSinkConfig object

Configuration for the splunk_hec_logs sink.

default_token string required

Wrapper for sensitive strings containing credentials

encoding required

Configures how events are encoded into raw bytes.

All of: object | object | object | object | object | object object | object | object | object | object | object, object object
endpoint string required

The scheme (http or https) must be specified. No path should be included since the paths defined by the Splunk API are used.

format=uri
acknowledgements

Splunk HEC acknowledgement configuration.

Default:
{
  "indexer_acknowledgements_enabled": true,
  "query_interval": 10,
  "retry_limit": 30,
  "max_pending_acks": 1000000
}
All of: object object, Controls how acknowledgements are handled for this sink. object
auto_extract_timestamp boolean | null

This option is only relevant to Splunk v8.x and above, and is only applied when endpoint_target is set to event.

Setting this to true causes Splunk to extract the timestamp from the message text rather than use the timestamp embedded in the event. The timestamp must be in the format yyyy-mm-dd hh:mm:ss.

Default: null

Event batching behavior.

Default:
{
  "max_bytes": null,
  "max_events": null,
  "timeout_secs": null
}
compression const: "none" | const: "gzip" | const: "zlib" | const: "zstd" | const: "snappy" | object

All compression algorithms use the default compression level unless otherwise specified.

endpoint_target const: "raw" | const: "event"

Splunk HEC endpoint configuration.

Default: "event"

By default, the global log_schema.host_key option is used if log events are Legacy namespaced, or the semantic meaning of "host" is used, if defined.

If not specified, the default index defined within Splunk is used.

Default:
[]
request object

Various settings can be configured, such as concurrency and rate limits, timeouts, retry behavior, etc.

Note that the retry backoff policy follows the Fibonacci sequence.

9 nested properties
adaptive_concurrency object

These parameters typically do not require changes from the default, and incorrect values can lead to meta-stable or unstable performance and sink behavior. Proceed with caution.

5 nested properties
decrease_ratio number

Valid values are greater than 0 and less than 1. Smaller values cause the algorithm to scale back rapidly when latency increases.

Note that the new limit is rounded down after applying this ratio.

Default: 0.9
min=0.0max=1.0
ewma_alpha number

Valid values are greater than 0 and less than 1.

ARC uses an exponentially weighted moving average (EWMA) of past RTT measurements as a reference to compare with the current RTT. Smaller values cause this reference to adjust more slowly, which may be useful if a service has unusually high response variability.

Default: 0.4
min=0.0max=1.0
initial_concurrency integer

It is recommended to set this value to your service's average limit if you're seeing that it takes a long time to ramp up adaptive concurrency after a restart. You can find this value by looking at the adaptive_concurrency_limit metric.

Default: 1
min=1.0max=9007199254740990.0
max_concurrency_limit integer

The adaptive request concurrency limit will not go above this bound. This is put in place as a safeguard.

Default: 200
min=1.0max=9007199254740990.0
rtt_deviation_scale number

Valid values are greater than or equal to 0, and we expect reasonable values to range from 1.0 to 3.0.

When calculating the past RTT average, we also compute a secondary “deviation” value that indicates how variable those values are. We use that deviation when comparing the past RTT average to the current measurements, so we can ignore increases in RTT that are within an expected range. This factor is used to scale up the deviation to an appropriate range. Larger values cause the algorithm to ignore larger increases in the RTT.

Default: 2.5
min=0.0max=9007199254740990.0
concurrency const: "none" | const: "adaptive" | integer

Configuration for outbound request concurrency.

This can be set either to one of the below enum values or to a positive integer, which denotes a fixed concurrency limit.

rate_limit_duration_secs integer

The time window used for the rate_limit_num option.

Default: 1
min=0.0max=9007199254740990.0
rate_limit_num integer

The maximum number of requests allowed within the rate_limit_duration_secs time window.

Default: 9223372036854775807
min=0.0max=9007199254740990.0
retry_attempts integer

The maximum number of retries to make for failed requests.

Default: 9223372036854775807
min=0.0max=9007199254740990.0
retry_initial_backoff_secs integer

After the first retry has failed, the fibonacci sequence is used to select future backoffs.

Default: 1
min=1.0max=9007199254740990.0
retry_jitter_mode const: "None" | const: "Full"

The jitter mode to use for retry backoff behavior.

retry_max_duration_secs integer

The maximum amount of time to wait between retries.

Default: 30
min=1.0max=9007199254740990.0
timeout_secs integer

Datadog highly recommends that you do not lower this value below the service's internal timeout, as this could create orphaned requests, pile on retries, and result in duplicate data downstream.

Default: 60
min=0.0max=9007199254740990.0

This is typically the filename the logs originated from.

If unset, the Splunk collector sets it.

If unset, Splunk defaults to httpevent.

By default, either the global log_schema.timestamp_key option is used if log events are Legacy namespaced, or the semantic meaning of "timestamp" is used, if defined.

vector::sinks::splunk_hec::metrics::config::HecMetricsSinkConfig object

Configuration of the splunk_hec_metrics sink.

default_token string required

Wrapper for sensitive strings containing credentials

endpoint string required

The scheme (http or https) must be specified. No path should be included since the paths defined by the Splunk API are used.

format=uri
acknowledgements

Splunk HEC acknowledgement configuration.

Default:
{
  "indexer_acknowledgements_enabled": true,
  "query_interval": 10,
  "retry_limit": 30,
  "max_pending_acks": 1000000
}
All of: object object, Controls how acknowledgements are handled for this sink. object

Event batching behavior.

Default:
{
  "max_bytes": null,
  "max_events": null,
  "timeout_secs": null
}
compression const: "none" | const: "gzip" | const: "zlib" | const: "zstd" | const: "snappy" | object

All compression algorithms use the default compression level unless otherwise specified.

default_namespace string | null

This namespace is only used if a metric has no existing namespace. When a namespace is present, it is used as a prefix to the metric name, and separated with a period (.).

host_key string

An optional path that deserializes an empty string to None.

If not specified, the default index defined within Splunk is used.

request object

Various settings can be configured, such as concurrency and rate limits, timeouts, retry behavior, etc.

Note that the retry backoff policy follows the Fibonacci sequence.

9 nested properties
adaptive_concurrency object

These parameters typically do not require changes from the default, and incorrect values can lead to meta-stable or unstable performance and sink behavior. Proceed with caution.

5 nested properties
decrease_ratio number

Valid values are greater than 0 and less than 1. Smaller values cause the algorithm to scale back rapidly when latency increases.

Note that the new limit is rounded down after applying this ratio.

Default: 0.9
min=0.0max=1.0
ewma_alpha number

Valid values are greater than 0 and less than 1.

ARC uses an exponentially weighted moving average (EWMA) of past RTT measurements as a reference to compare with the current RTT. Smaller values cause this reference to adjust more slowly, which may be useful if a service has unusually high response variability.

Default: 0.4
min=0.0max=1.0
initial_concurrency integer

It is recommended to set this value to your service's average limit if you're seeing that it takes a long time to ramp up adaptive concurrency after a restart. You can find this value by looking at the adaptive_concurrency_limit metric.

Default: 1
min=1.0max=9007199254740990.0
max_concurrency_limit integer

The adaptive request concurrency limit will not go above this bound. This is put in place as a safeguard.

Default: 200
min=1.0max=9007199254740990.0
rtt_deviation_scale number

Valid values are greater than or equal to 0, and we expect reasonable values to range from 1.0 to 3.0.

When calculating the past RTT average, we also compute a secondary “deviation” value that indicates how variable those values are. We use that deviation when comparing the past RTT average to the current measurements, so we can ignore increases in RTT that are within an expected range. This factor is used to scale up the deviation to an appropriate range. Larger values cause the algorithm to ignore larger increases in the RTT.

Default: 2.5
min=0.0max=9007199254740990.0
concurrency const: "none" | const: "adaptive" | integer

Configuration for outbound request concurrency.

This can be set either to one of the below enum values or to a positive integer, which denotes a fixed concurrency limit.

rate_limit_duration_secs integer

The time window used for the rate_limit_num option.

Default: 1
min=0.0max=9007199254740990.0
rate_limit_num integer

The maximum number of requests allowed within the rate_limit_duration_secs time window.

Default: 9223372036854775807
min=0.0max=9007199254740990.0
retry_attempts integer

The maximum number of retries to make for failed requests.

Default: 9223372036854775807
min=0.0max=9007199254740990.0
retry_initial_backoff_secs integer

After the first retry has failed, the fibonacci sequence is used to select future backoffs.

Default: 1
min=1.0max=9007199254740990.0
retry_jitter_mode const: "None" | const: "Full"

The jitter mode to use for retry backoff behavior.

retry_max_duration_secs integer

The maximum amount of time to wait between retries.

Default: 30
min=1.0max=9007199254740990.0
timeout_secs integer

Datadog highly recommends that you do not lower this value below the service's internal timeout, as this could create orphaned requests, pile on retries, and result in duplicate data downstream.

Default: 60
min=0.0max=9007199254740990.0

This is typically the filename the logs originated from.

If unset, the Splunk collector sets it.

If unset, Splunk defaults to httpevent.

vector::sinks::statsd::config::StatsdSinkConfig

Configuration for the statsd sink.

vector::sinks::util::adaptive_concurrency::AdaptiveConcurrencySettings object

These parameters typically do not require changes from the default, and incorrect values can lead to meta-stable or unstable performance and sink behavior. Proceed with caution.

decrease_ratio number

Valid values are greater than 0 and less than 1. Smaller values cause the algorithm to scale back rapidly when latency increases.

Note that the new limit is rounded down after applying this ratio.

Default: 0.9
min=0.0max=1.0
ewma_alpha number

Valid values are greater than 0 and less than 1.

ARC uses an exponentially weighted moving average (EWMA) of past RTT measurements as a reference to compare with the current RTT. Smaller values cause this reference to adjust more slowly, which may be useful if a service has unusually high response variability.

Default: 0.4
min=0.0max=1.0
initial_concurrency integer

It is recommended to set this value to your service's average limit if you're seeing that it takes a long time to ramp up adaptive concurrency after a restart. You can find this value by looking at the adaptive_concurrency_limit metric.

Default: 1
min=1.0max=9007199254740990.0
max_concurrency_limit integer

The adaptive request concurrency limit will not go above this bound. This is put in place as a safeguard.

Default: 200
min=1.0max=9007199254740990.0
rtt_deviation_scale number

Valid values are greater than or equal to 0, and we expect reasonable values to range from 1.0 to 3.0.

When calculating the past RTT average, we also compute a secondary “deviation” value that indicates how variable those values are. We use that deviation when comparing the past RTT average to the current measurements, so we can ignore increases in RTT that are within an expected range. This factor is used to scale up the deviation to an appropriate range. Larger values cause the algorithm to ignore larger increases in the RTT.

Default: 2.5
min=0.0max=9007199254740990.0
vector::sinks::util::batch::BatchConfig<vector::sinks::greptimedb::GreptimeDBDefaultBatchSettings> object

Event batching behavior.

max_bytes integer | null

This is based on the uncompressed size of the batched events, before they are serialized/compressed.

Default: null
min=0.0max=9007199254740990.0
max_events integer | null

The maximum size of a batch before it is flushed.

Default: 20
min=0.0max=9007199254740990.0
timeout_secs number | null

The maximum age of a batch before it is flushed.

Default: 1.0
min=-9007199254740990.0max=9007199254740990.0
vector::sinks::util::batch::BatchConfig<vector::sinks::splunk_hec::common::util::SplunkHecDefaultBatchSettings> object

Event batching behavior.

max_bytes integer | null

This is based on the uncompressed size of the batched events, before they are serialized/compressed.

Default: 1000000
min=0.0max=9007199254740990.0
max_events integer | null

The maximum size of a batch before it is flushed.

Default: null
min=0.0max=9007199254740990.0
timeout_secs number | null

The maximum age of a batch before it is flushed.

Default: 1.0
min=-9007199254740990.0max=9007199254740990.0
vector::sinks::util::batch::BatchConfig<vector::sinks::util::batch::BulkSizeBasedDefaultBatchSettings> object

Event batching behavior.

max_bytes integer | null

This is based on the uncompressed size of the batched events, before they are serialized/compressed.

Default: 10000000
min=0.0max=9007199254740990.0
max_events integer | null

The maximum size of a batch before it is flushed.

Default: null
min=0.0max=9007199254740990.0
timeout_secs number | null

The maximum age of a batch before it is flushed.

Default: 300.0
min=-9007199254740990.0max=9007199254740990.0
vector::sinks::util::batch::BatchConfig<vector::sinks::util::batch::RealtimeSizeBasedDefaultBatchSettings> object

Event batching behavior.

max_bytes integer | null

This is based on the uncompressed size of the batched events, before they are serialized/compressed.

Default: 10000000
min=0.0max=9007199254740990.0
max_events integer | null

The maximum size of a batch before it is flushed.

Default: null
min=0.0max=9007199254740990.0
timeout_secs number | null

The maximum age of a batch before it is flushed.

Default: 1.0
min=-9007199254740990.0max=9007199254740990.0
vector::sinks::util::buffer::compression::Compression const: "none" | const: "gzip" | const: "zlib" | const: "zstd" | const: "snappy" | object

All compression algorithms use the default compression level unless otherwise specified.

vector::sinks::util::buffer::compression::CompressionLevel enum

Compression level.

vector::sinks::util::http::RequestConfig

Outbound HTTP request settings.

vector::sinks::util::retries::JitterMode const: "None" | const: "Full"

The jitter mode to use for retry backoff behavior.

vector::sinks::util::service::TowerRequestConfig object

Various settings can be configured, such as concurrency and rate limits, timeouts, retry behavior, etc.

Note that the retry backoff policy follows the Fibonacci sequence.

adaptive_concurrency object

These parameters typically do not require changes from the default, and incorrect values can lead to meta-stable or unstable performance and sink behavior. Proceed with caution.

5 nested properties
decrease_ratio number

Valid values are greater than 0 and less than 1. Smaller values cause the algorithm to scale back rapidly when latency increases.

Note that the new limit is rounded down after applying this ratio.

Default: 0.9
min=0.0max=1.0
ewma_alpha number

Valid values are greater than 0 and less than 1.

ARC uses an exponentially weighted moving average (EWMA) of past RTT measurements as a reference to compare with the current RTT. Smaller values cause this reference to adjust more slowly, which may be useful if a service has unusually high response variability.

Default: 0.4
min=0.0max=1.0
initial_concurrency integer

It is recommended to set this value to your service's average limit if you're seeing that it takes a long time to ramp up adaptive concurrency after a restart. You can find this value by looking at the adaptive_concurrency_limit metric.

Default: 1
min=1.0max=9007199254740990.0
max_concurrency_limit integer

The adaptive request concurrency limit will not go above this bound. This is put in place as a safeguard.

Default: 200
min=1.0max=9007199254740990.0
rtt_deviation_scale number

Valid values are greater than or equal to 0, and we expect reasonable values to range from 1.0 to 3.0.

When calculating the past RTT average, we also compute a secondary “deviation” value that indicates how variable those values are. We use that deviation when comparing the past RTT average to the current measurements, so we can ignore increases in RTT that are within an expected range. This factor is used to scale up the deviation to an appropriate range. Larger values cause the algorithm to ignore larger increases in the RTT.

Default: 2.5
min=0.0max=9007199254740990.0
concurrency const: "none" | const: "adaptive" | integer

Configuration for outbound request concurrency.

This can be set either to one of the below enum values or to a positive integer, which denotes a fixed concurrency limit.

rate_limit_duration_secs integer

The time window used for the rate_limit_num option.

Default: 1
min=0.0max=9007199254740990.0
rate_limit_num integer

The maximum number of requests allowed within the rate_limit_duration_secs time window.

Default: 9223372036854775807
min=0.0max=9007199254740990.0
retry_attempts integer

The maximum number of retries to make for failed requests.

Default: 9223372036854775807
min=0.0max=9007199254740990.0
retry_initial_backoff_secs integer

After the first retry has failed, the fibonacci sequence is used to select future backoffs.

Default: 1
min=1.0max=9007199254740990.0
retry_jitter_mode const: "None" | const: "Full"

The jitter mode to use for retry backoff behavior.

retry_max_duration_secs integer

The maximum amount of time to wait between retries.

Default: 30
min=1.0max=9007199254740990.0
timeout_secs integer

Datadog highly recommends that you do not lower this value below the service's internal timeout, as this could create orphaned requests, pile on retries, and result in duplicate data downstream.

Default: 60
min=0.0max=9007199254740990.0
vector::sinks::util::service::concurrency::Concurrency const: "none" | const: "adaptive" | integer

Configuration for outbound request concurrency.

This can be set either to one of the below enum values or to a positive integer, which denotes a fixed concurrency limit.

vector::sinks::util::service::net::HostAndPort string

Both IP addresses and hostnames/fully qualified domain names (FQDNs) are accepted formats.

The address must include a port.

vector::sinks::util::uri::UriSerde string
vector::sinks::vector::config::VectorConfig object

Configuration for the vector sink.

address string required

Both IP address and hostname are accepted formats.

The address must include a port.

format=uri
acknowledgements object

See End-to-end Acknowledgements for more information on how event acknowledgement is handled.

1 nested properties
enabled boolean | null

When enabled for a sink, any source connected to that sink, where the source supports end-to-end acknowledgements as well, waits for events to be acknowledged by all connected sinks before acknowledging them at the source.

Enabling or disabling acknowledgements at the sink level takes precedence over any global acknowledgements configuration.

batch object

Event batching behavior.

Default:
{
  "max_bytes": null,
  "max_events": null,
  "timeout_secs": null
}
3 nested properties
max_bytes integer | null

This is based on the uncompressed size of the batched events, before they are serialized/compressed.

Default: null
min=0.0max=9007199254740990.0
max_events integer | null

The maximum size of a batch before it is flushed.

Default: 1000
min=0.0max=9007199254740990.0
timeout_secs number | null

The maximum age of a batch before it is flushed.

Default: 1.0
min=-9007199254740990.0max=9007199254740990.0
compression boolean

If set to true, requests are compressed with gzip.

Default: false
request object

Various settings can be configured, such as concurrency and rate limits, timeouts, retry behavior, etc.

Note that the retry backoff policy follows the Fibonacci sequence.

9 nested properties
adaptive_concurrency object

These parameters typically do not require changes from the default, and incorrect values can lead to meta-stable or unstable performance and sink behavior. Proceed with caution.

5 nested properties
decrease_ratio number

Valid values are greater than 0 and less than 1. Smaller values cause the algorithm to scale back rapidly when latency increases.

Note that the new limit is rounded down after applying this ratio.

Default: 0.9
min=0.0max=1.0
ewma_alpha number

Valid values are greater than 0 and less than 1.

ARC uses an exponentially weighted moving average (EWMA) of past RTT measurements as a reference to compare with the current RTT. Smaller values cause this reference to adjust more slowly, which may be useful if a service has unusually high response variability.

Default: 0.4
min=0.0max=1.0
initial_concurrency integer

It is recommended to set this value to your service's average limit if you're seeing that it takes a long time to ramp up adaptive concurrency after a restart. You can find this value by looking at the adaptive_concurrency_limit metric.

Default: 1
min=1.0max=9007199254740990.0
max_concurrency_limit integer

The adaptive request concurrency limit will not go above this bound. This is put in place as a safeguard.

Default: 200
min=1.0max=9007199254740990.0
rtt_deviation_scale number

Valid values are greater than or equal to 0, and we expect reasonable values to range from 1.0 to 3.0.

When calculating the past RTT average, we also compute a secondary “deviation” value that indicates how variable those values are. We use that deviation when comparing the past RTT average to the current measurements, so we can ignore increases in RTT that are within an expected range. This factor is used to scale up the deviation to an appropriate range. Larger values cause the algorithm to ignore larger increases in the RTT.

Default: 2.5
min=0.0max=9007199254740990.0
concurrency const: "none" | const: "adaptive" | integer

Configuration for outbound request concurrency.

This can be set either to one of the below enum values or to a positive integer, which denotes a fixed concurrency limit.

rate_limit_duration_secs integer

The time window used for the rate_limit_num option.

Default: 1
min=0.0max=9007199254740990.0
rate_limit_num integer

The maximum number of requests allowed within the rate_limit_duration_secs time window.

Default: 9223372036854775807
min=0.0max=9007199254740990.0
retry_attempts integer

The maximum number of retries to make for failed requests.

Default: 9223372036854775807
min=0.0max=9007199254740990.0
retry_initial_backoff_secs integer

After the first retry has failed, the fibonacci sequence is used to select future backoffs.

Default: 1
min=1.0max=9007199254740990.0
retry_jitter_mode const: "None" | const: "Full"

The jitter mode to use for retry backoff behavior.

retry_max_duration_secs integer

The maximum amount of time to wait between retries.

Default: 30
min=1.0max=9007199254740990.0
timeout_secs integer

Datadog highly recommends that you do not lower this value below the service's internal timeout, as this could create orphaned requests, pile on retries, and result in duplicate data downstream.

Default: 60
min=0.0max=9007199254740990.0
version null | const: "2"

Version of the configuration.

vector::sinks::webhdfs::config::WebHdfsConfig

Configuration for the webhdfs sink.

vector::sinks::websocket::config::WebSocketSinkConfig object

Configuration for the websocket sink.

encoding required

Configures how events are encoded into raw bytes.

All of: object | object | object | object | object | object object | object | object | object | object | object, object object
uri string required

This should include the protocol and host, but can also include the port, path, and any other valid part of a URI.

acknowledgements object

See End-to-end Acknowledgements for more information on how event acknowledgement is handled.

1 nested properties
enabled boolean | null

When enabled for a sink, any source connected to that sink, where the source supports end-to-end acknowledgements as well, waits for events to be acknowledged by all connected sinks before acknowledging them at the source.

Enabling or disabling acknowledgements at the sink level takes precedence over any global acknowledgements configuration.

ping_interval integer | null

If this option is not configured, pings are not sent on an interval.

If the ping_timeout is not set, pings are still sent but there is no expectation of pong response times.

min=1.0max=9007199254740990.0
ping_timeout integer | null

If a response is not received within this time, the connection is re-established.

min=1.0max=9007199254740990.0
vector::sources::Sources

Configurable sources in Vector.

vector::sources::amqp::AmqpSourceConfig

Supports AMQP version 0.9.1

vector::sources::apache_metrics::ApacheMetricsConfig object

Configuration for the apache_metrics source.

endpoints string[] required

The list of mod_status endpoints to scrape metrics from.

namespace string

Disabled if empty.

Default: "apache"
scrape_interval_secs integer

A span of time, in whole seconds.

min=0.0max=9007199254740990.0
vector::sources::aws_ecs_metrics::AwsEcsMetricsSourceConfig object

Configuration for the aws_ecs_metrics source.

endpoint string

If empty, the URI is automatically discovered based on the latest version detected.

By default:

  • The version 4 endpoint base URI is stored in the environment variable ECS_CONTAINER_METADATA_URI_V4.
  • The version 3 endpoint base URI is stored in the environment variable ECS_CONTAINER_METADATA_URI.
  • The version 2 endpoint base URI is 169.254.170.2/v2/.
Default: "http://169.254.170.2/v2"
namespace string

Disabled if empty.

Default: "awsecs"
scrape_interval_secs integer

A span of time, in whole seconds.

min=0.0max=9007199254740990.0
version const: "v2" | const: "v3" | const: "v4"

If empty, the version is automatically discovered based on environment variables.

By default:

  • Version 4 is used if the environment variable ECS_CONTAINER_METADATA_URI_V4 is defined.
  • Version 3 is used if the environment variable ECS_CONTAINER_METADATA_URI_V4 is not defined, but the environment variable ECS_CONTAINER_METADATA_URI is defined.
  • Version 2 is used if neither of the environment variables ECS_CONTAINER_METADATA_URI_V4 or ECS_CONTAINER_METADATA_URI are defined.
Default: "v2"
vector::sources::aws_kinesis_firehose::AwsKinesisFirehoseConfig object

Configuration for the aws_kinesis_firehose source.

address string required

An internet socket address, either IPv4 or IPv6.

store_access_key boolean required

If set to true, when incoming requests contains an access key sent by AWS Firehose, it is kept in the event secrets as "aws_kinesis_firehose_access_key".

AWS Kinesis Firehose can be configured to pass along a user-configurable access key with each request. If configured, access_key should be set to the same value. Otherwise, all requests are allowed.

access_keys array | null

AWS Kinesis Firehose can be configured to pass along a user-configurable access key with each request. If configured, access_keys should be set to the same value. Otherwise, all requests are allowed.

acknowledgements object

This setting is deprecated in favor of enabling acknowledgements at the global or sink level.

Enabling or disabling acknowledgements at the source level has no effect on acknowledgement behavior.

See End-to-end Acknowledgements for more information on how event acknowledgement is handled.

1 nested properties
enabled boolean | null

Whether or not end-to-end acknowledgements are enabled for this source.

decoding object | object | object

Configures how events are decoded from raw bytes.

framing object

Framing handles how events are separated when encoded in a raw byte form, where each event is a frame that must be prefixed, or delimited, in a way that marks where an event begins and ends within the byte stream.

keepalive object

Configuration of HTTP server keepalive parameters.

2 nested properties
max_connection_age_jitter_factor number

A value of 0.1 means that the actual duration will be between 90% and 110% of the specified maximum duration.

Default: 0.1
min=0.0max=1.0
max_connection_age_secs integer | null

Only applies to HTTP/0.9, HTTP/1.0, and HTTP/1.1 requests.

A random jitter configured by max_connection_age_jitter_factor is added to the specified duration to spread out connection storms.

Default: 300
min=0.0max=9007199254740990.0
log_namespace boolean | null

The namespace to use for logs. This overrides the global setting.

Default: null
record_compression const: "auto" | const: "none" | const: "gzip"

Some services, like AWS CloudWatch Logs, compresses the events with gzip, before sending them AWS Kinesis Firehose. This option can be used to automatically decompress them before forwarding them to the next component.

Note that this is different from Content encoding option of the Firehose HTTP endpoint destination. That option controls the content encoding of the entire HTTP request.

Default: "auto"
vector::sources::aws_s3::AwsS3Config

Configuration for the aws_s3 source.

vector::sources::aws_sqs::config::AwsSqsConfig

Configuration for the aws_sqs source.

vector::sources::datadog_agent::DatadogAgentConfig object

Configuration for the datadog_agent source.

address string required

An internet socket address, either IPv4 or IPv6.

acknowledgements object

This setting is deprecated in favor of enabling acknowledgements at the global or sink level.

Enabling or disabling acknowledgements at the source level has no effect on acknowledgement behavior.

See End-to-end Acknowledgements for more information on how event acknowledgement is handled.

1 nested properties
enabled boolean | null

Whether or not end-to-end acknowledgements are enabled for this source.

decoding object | object | object

Configures how events are decoded from raw bytes.

disable_logs boolean

If this is set to true, logs are not accepted by the component.

Default: false
disable_metrics boolean

If this is set to true, metrics (beta) are not accepted by the component.

Default: false
disable_traces boolean

If this is set to true, traces (alpha) are not accepted by the component.

Default: false
framing object

Framing handles how events are separated when encoded in a raw byte form, where each event is a frame that must be prefixed, or delimited, in a way that marks where an event begins and ends within the byte stream.

keepalive object

Configuration of HTTP server keepalive parameters.

2 nested properties
max_connection_age_jitter_factor number

A value of 0.1 means that the actual duration will be between 90% and 110% of the specified maximum duration.

Default: 0.1
min=0.0max=1.0
max_connection_age_secs integer | null

Only applies to HTTP/0.9, HTTP/1.0, and HTTP/1.1 requests.

A random jitter configured by max_connection_age_jitter_factor is added to the specified duration to spread out connection storms.

Default: 300
min=0.0max=9007199254740990.0
log_namespace boolean | null

The namespace to use for logs. This overrides the global setting.

Default: null
multiple_outputs boolean

For a source component named agent, the received logs, metrics (beta), and traces (alpha) can then be configured as input to other components by specifying agent.logs, agent.metrics, and agent.traces, respectively.

Default: false
parse_ddtags boolean

If this is set to true, when log events contain the field ddtags, the string value that contains a list of key:value pairs set by the Agent is parsed and expanded into an array.

Default: false
store_api_key boolean

If this is set to true, when incoming events contain a Datadog API key, it is stored in the event metadata and used if the event is sent to a Datadog sink.

Default: true
vector::sources::demo_logs::DemoLogsConfig

Configuration for the demo_logs source.

vector::sources::dnstap::DnstapConfig

Configuration for the dnstap source.

vector::sources::docker_logs::DockerLogsConfig object

Configuration for the docker_logs source.

auto_partial_merge boolean

Enables automatic merging of partial events.

docker_host string | null

Use an HTTPS URL to enable TLS encryption.

If absent, the DOCKER_HOST environment variable is used. If DOCKER_HOST is also absent, the default Docker local socket (/var/run/docker.sock on Unix platforms, //./pipe/docker_engine on Windows) is used.

exclude_containers array | null

Matching is prefix first, so specifying a value of foo would match any container named foo as well as any container whose name started with foo. This applies equally whether matching container IDs or names.

By default, the source collects logs for all containers. If exclude_containers is configured, any container that matches a configured exclusion is excluded even if it is also included with include_containers, so care should be taken when using prefix matches as they cannot be overridden by a corresponding entry in include_containers, for example, excluding foo by attempting to include foo-specific-id.

This can be used in conjunction with include_containers.

include_containers array | null

Matching is prefix first, so specifying a value of foo would match any container named foo as well as any container whose name started with foo. This applies equally whether matching container IDs or names.

By default, the source collects logs for all containers. If include_containers is configured, only containers that match a configured inclusion and are also not excluded get matched.

This can be used in conjunction with exclude_containers.

include_images array | null

If not provided, all images are included.

include_labels array | null

Labels should follow the syntax described in the Docker object labels documentation.

log_namespace boolean | null

The namespace to use for logs. This overrides the global setting.

Default: null

If not specified, multiline aggregation is disabled.

partial_event_marker_field string | null

If auto_partial_merge is disabled, partial events are emitted with a log field, set by this configuration value, indicating that the event is not complete.

Default: "_partial"
retry_backoff_secs integer

A span of time, in whole seconds.

min=0.0max=9007199254740990.0
tls null | object

Only relevant when connecting to Docker with an HTTPS URL.

If not configured, the environment variable DOCKER_CERT_PATH is used. If DOCKER_CERT_PATH is absent, then DOCKER_CONFIG is used. If both environment variables are absent, the certificates in ~/.docker/ are read.

vector::sources::eventstoredb_metrics::EventStoreDbConfig object

Configuration for the eventstoredb_metrics source.

default_namespace string | null

By default, eventstoredb is used.

endpoint string

Endpoint to scrape stats from.

Default: "https://localhost:2113/stats"
scrape_interval_secs integer

A span of time, in whole seconds.

min=0.0max=9007199254740990.0
vector::sources::exec::ExecConfig object

Configuration for the exec source.

command string[] required

The command to run, plus any arguments required.

mode const: "scheduled" | const: "streaming" required

Mode of operation for running the command.

clear_environment boolean

Whether or not to clear the environment before setting custom environment variables.

Default: false
decoding object | object | object

Configures how events are decoded from raw bytes.

environment Record<string, string>

Custom environment variables to set or update when running the command. If a variable name already exists in the environment, its value is replaced.

Default: null
include_stderr boolean

Whether or not the output from stderr should be included when generating events.

Default: true
log_namespace boolean | null

The namespace to use for logs. This overrides the global setting.

Default: null
maximum_buffer_size_bytes integer

The maximum buffer size allowed before a log event is generated.

Default: 1000000
min=0.0max=9007199254740990.0
scheduled null | object

Configuration options for scheduled commands.

streaming null | object

Configuration options for streaming commands.

The directory in which to run the command.

vector::sources::file::FileConfig object

Configuration for the file source.

include stdlib::PathBuf[] required

Array of file patterns to include. Globbing is supported.

acknowledgements object

This setting is deprecated in favor of enabling acknowledgements at the global or sink level.

Enabling or disabling acknowledgements at the source level has no effect on acknowledgement behavior.

See End-to-end Acknowledgements for more information on how event acknowledgement is handled.

1 nested properties
enabled boolean | null

Whether or not end-to-end acknowledgements are enabled for this source.

By default, the global data_dir option is used. Make sure the running user has write permissions to this directory.

If this directory is specified, then Vector will attempt to create it.

Default: null
encoding null | object

Character set encoding.

Default: null

Takes precedence over the include option. Note: The exclude patterns are applied after the attempt to glob everything in include. This means that all files are first matched by include and then filtered by the exclude patterns. This can be impactful if include contains directories with contents that are not accessible.

Default:
[]
file_key string

An optional path that deserializes an empty string to None.

fingerprint object | object

This is important for checkpointing when file rotation is used.

Default:
{
  "strategy": "checksum",
  "bytes": null,
  "ignored_header_bytes": 0,
  "lines": 1
}
glob_minimum_cooldown_ms integer

A span of time, in whole milliseconds.

min=0.0max=9007199254740990.0
ignore_checkpoints boolean | null

Checkpoints are still written normally.

Default: null
ignore_not_found boolean

This may be useful when used with source directories containing dangling symlinks.

Default: false
ignore_older_secs integer | null

Ignore files with a data modification date older than the specified number of seconds.

Default: null
min=0.0max=9007199254740990.0
internal_metrics object

Configuration of internal metrics for file-based components.

1 nested properties
include_file_tag boolean

This is useful for distinguishing between different files while monitoring. However, the tag's cardinality is unbounded.

Default: false
line_delimiter string

String sequence used to separate one file line from another.

Default: " "
log_namespace boolean | null

The namespace to use for logs. This overrides the global setting.

Default: null
max_line_bytes integer

This protects against malformed lines or tailing incorrect files.

Default: 102400
min=0.0max=9007199254740990.0
max_read_bytes integer

This allows distributing the reads more or less evenly across the files.

Default: 2048
min=0.0max=9007199254740990.0
message_start_indicator string | null

String value used to identify the start of a multi-line message.

Default: null
multi_line_timeout integer

How long to wait for more data when aggregating a multi-line message, in milliseconds.

Default: 1000
min=0.0max=9007199254740990.0

If not specified, multiline aggregation is disabled.

Default: null

The value is the byte offset of the start of the line within the file.

Off by default, the offset is only added to the event if this is set.

Default: null
oldest_first boolean

Instead of balancing read capacity fairly across all watched files, prioritize draining the oldest files before moving on to read data from more recent files.

Default: false
read_from const: "beginning" | const: "end"

File position to use when reading a new file.

remove_after_secs integer | null

If not specified, files are not removed.

Default: null
min=0.0max=9007199254740990.0
rotate_wait_secs integer

A span of time, in whole seconds.

min=0.0max=9007199254740990.0
start_at_beginning boolean | null

Whether or not to start reading from the beginning of a new file.

Default: null
vector::sources::file_descriptors::file_descriptor::FileDescriptorSourceConfig object

Configuration for the file_descriptor source.

fd integer required

The file descriptor number to read from.

min=0.0max=4294967295.0
decoding object | object | object

Configures how events are decoded from raw bytes.

log_namespace boolean | null

The namespace to use for logs. This overrides the global setting.

Default: null
max_length integer

Messages larger than this are truncated.

Default: 102400
min=0.0max=9007199254740990.0
vector::sources::file_descriptors::stdin::StdinConfig object

Configuration for the stdin source.

decoding object | object | object

Configures how events are decoded from raw bytes.

log_namespace boolean | null

The namespace to use for logs. This overrides the global setting.

Default: null
max_length integer

Messages larger than this are truncated.

Default: 102400
min=0.0max=9007199254740990.0
vector::sources::fluent::FluentConfig object

Configuration for the fluent source.

address string required

If a socket address is used, it must include a port.

acknowledgements object

This setting is deprecated in favor of enabling acknowledgements at the global or sink level.

Enabling or disabling acknowledgements at the source level has no effect on acknowledgement behavior.

See End-to-end Acknowledgements for more information on how event acknowledgement is handled.

1 nested properties
enabled boolean | null

Whether or not end-to-end acknowledgements are enabled for this source.

connection_limit integer | null

The maximum number of TCP connections that are allowed at any given time.

min=0.0max=4294967295.0
log_namespace boolean | null

The namespace to use for logs. This overrides the global setting.

Default: null
receive_buffer_bytes integer | null

This generally should not need to be changed.

min=0.0max=9007199254740990.0
vector::sources::gcp_pubsub::PubsubConfig

Configuration for the gcp_pubsub source.

vector::sources::heroku_logs::LogplexConfig object

Configuration for heroku_logs source.

address string required

An internet socket address, either IPv4 or IPv6.

acknowledgements object

This setting is deprecated in favor of enabling acknowledgements at the global or sink level.

Enabling or disabling acknowledgements at the source level has no effect on acknowledgement behavior.

See End-to-end Acknowledgements for more information on how event acknowledgement is handled.

1 nested properties
enabled boolean | null

Whether or not end-to-end acknowledgements are enabled for this source.

decoding object | object | object

Configures how events are decoded from raw bytes.

framing object

Framing handles how events are separated when encoded in a raw byte form, where each event is a frame that must be prefixed, or delimited, in a way that marks where an event begins and ends within the byte stream.

keepalive object

Configuration of HTTP server keepalive parameters.

2 nested properties
max_connection_age_jitter_factor number

A value of 0.1 means that the actual duration will be between 90% and 110% of the specified maximum duration.

Default: 0.1
min=0.0max=1.0
max_connection_age_secs integer | null

Only applies to HTTP/0.9, HTTP/1.0, and HTTP/1.1 requests.

A random jitter configured by max_connection_age_jitter_factor is added to the specified duration to spread out connection storms.

Default: 300
min=0.0max=9007199254740990.0
log_namespace boolean | null

The namespace to use for logs. This overrides the global setting.

Default: null
query_parameters string[]

Accepts the wildcard (*) character for query parameters matching a specified pattern.

Specifying "*" results in all query parameters included in the log event.

These override any values included in the body with conflicting names.

Default:
[]
vector::sources::host_metrics::FilterList object

Filtering configuration.

excludes array | null

The patterns are matched using globbing.

includes array | null

The patterns are matched using globbing.

vector::sources::host_metrics::HostMetricsConfig object

Configuration for the host_metrics source.

cgroups null | object

This collector is only available on Linux systems, and only supports either version 2 or hybrid cgroups.

Default:
{
  "levels": 100,
  "base": null,
  "groups": {
    "includes": null,
    "excludes": null
  }
}
collectors array | null

Defaults to all collectors.

Default:
[
  "cpu",
  "disk",
  "filesystem",
  "load",
  "host",
  "memory",
  "network",
  "process",
  "cgroups"
]
disk object

Options for the disk metrics collector.

Default:
{
  "devices": {
    "includes": null,
    "excludes": null
  }
}
1 nested properties
devices object

Filtering configuration.

2 nested properties
excludes array | null

The patterns are matched using globbing.

includes array | null

The patterns are matched using globbing.

filesystem object

Options for the filesystem metrics collector.

Default:
{
  "devices": {
    "includes": null,
    "excludes": null
  },
  "filesystems": {
    "includes": null,
    "excludes": null
  },
  "mountpoints": {
    "includes": null,
    "excludes": null
  }
}
3 nested properties
devices object

Filtering configuration.

2 nested properties
excludes array | null

The patterns are matched using globbing.

includes array | null

The patterns are matched using globbing.

filesystems object

Filtering configuration.

2 nested properties
excludes array | null

The patterns are matched using globbing.

includes array | null

The patterns are matched using globbing.

mountpoints object

Filtering configuration.

2 nested properties
excludes array | null

The patterns are matched using globbing.

includes array | null

The patterns are matched using globbing.

namespace string | null

Overrides the default namespace for the metrics emitted by the source.

Default: "host"
network object

Options for the network metrics collector.

Default:
{
  "devices": {
    "includes": null,
    "excludes": null
  }
}
1 nested properties
devices object

Filtering configuration.

2 nested properties
excludes array | null

The patterns are matched using globbing.

includes array | null

The patterns are matched using globbing.

process object

Options for the process metrics collector.

Default:
{
  "processes": {
    "includes": null,
    "excludes": null
  }
}
1 nested properties
processes object

Filtering configuration.

2 nested properties
excludes array | null

The patterns are matched using globbing.

includes array | null

The patterns are matched using globbing.

scrape_interval_secs integer

A span of time, in whole seconds.

min=0.0max=9007199254740990.0
vector::sources::http_client::client::HttpClientConfig object

Configuration for the http_client source.

endpoint string required

The full path must be specified.

decoding object | object | object

Configures how events are decoded from raw bytes.

framing object

Framing handles how events are separated when encoded in a raw byte form, where each event is a frame that must be prefixed, or delimited, in a way that marks where an event begins and ends within the byte stream.

headers Record<string, string[]>

One or more values for the same header can be provided.

Default:
{}
log_namespace boolean | null

The namespace to use for logs. This overrides the global setting.

Default: null
method const: "HEAD" | const: "GET" | const: "POST" | const: "PUT" | const: "PATCH" | const: "DELETE" | const: "OPTIONS"

HTTP method.

query Record<string, string[]>

One or more values for the same parameter key can be provided.

The parameters provided in this option are appended to any parameters manually provided in the endpoint option.

Default:
{}
scrape_interval_secs integer

A span of time, in whole seconds.

min=0.0max=9007199254740990.0
scrape_timeout_secs number

A span of time, in fractional seconds.

min=-9007199254740990.0max=9007199254740990.0
vector::sources::http_server::HttpConfig object

Configuration for the http_server source.

address string required

An internet socket address, either IPv4 or IPv6.

acknowledgements object

This setting is deprecated in favor of enabling acknowledgements at the global or sink level.

Enabling or disabling acknowledgements at the source level has no effect on acknowledgement behavior.

See End-to-end Acknowledgements for more information on how event acknowledgement is handled.

1 nested properties
enabled boolean | null

Whether or not end-to-end acknowledgements are enabled for this source.

decoding null | object | object | object

Configures how events are decoded from raw bytes.

encoding null | const: "text" | const: "ndjson" | const: "json" | const: "binary"

For json and ndjson encodings, the fields of the JSON objects are output as separate fields.

Default: null
headers string[]

Accepts the wildcard (*) character for headers matching a specified pattern.

Specifying "*" results in all headers included in the log event.

These headers are not included in the JSON payload if a field with a conflicting name exists.

Default:
[]
host_key string

An optional path that deserializes an empty string to None.

keepalive object

Configuration of HTTP server keepalive parameters.

2 nested properties
max_connection_age_jitter_factor number

A value of 0.1 means that the actual duration will be between 90% and 110% of the specified maximum duration.

Default: 0.1
min=0.0max=1.0
max_connection_age_secs integer | null

Only applies to HTTP/0.9, HTTP/1.0, and HTTP/1.1 requests.

A random jitter configured by max_connection_age_jitter_factor is added to the specified duration to spread out connection storms.

Default: 300
min=0.0max=9007199254740990.0
log_namespace boolean | null

The namespace to use for logs. This overrides the global setting.

Default: null
method const: "HEAD" | const: "GET" | const: "POST" | const: "PUT" | const: "PATCH" | const: "DELETE" | const: "OPTIONS"

HTTP method.

path string

The URL path on which log event POST requests are sent.

Default: "/"
path_key string

An optional path that deserializes an empty string to None.

query_parameters string[]

Accepts the wildcard (*) character for query parameters matching a specified pattern.

Specifying "*" results in all query parameters included in the log event.

These override any values included in the body with conflicting names.

Default:
[]
response_code integer

Specifies the HTTP response status code that will be returned on successful requests.

Default: 200
min=0.0max=65535.0
strict_path boolean

If set to true, only requests using the exact URL path specified in path are accepted. Otherwise, requests sent to a URL path that starts with the value of path are accepted.

With strict_path set to false and path set to "", the configured HTTP source accepts requests from any URL path.

Default: true
vector::sources::http_server::SimpleHttpConfig object

Configuration for the http_server source.

address string required

An internet socket address, either IPv4 or IPv6.

acknowledgements object

This setting is deprecated in favor of enabling acknowledgements at the global or sink level.

Enabling or disabling acknowledgements at the source level has no effect on acknowledgement behavior.

See End-to-end Acknowledgements for more information on how event acknowledgement is handled.

1 nested properties
enabled boolean | null

Whether or not end-to-end acknowledgements are enabled for this source.

decoding null | object | object | object

Configures how events are decoded from raw bytes.

encoding null | const: "text" | const: "ndjson" | const: "json" | const: "binary"

For json and ndjson encodings, the fields of the JSON objects are output as separate fields.

Default: null
headers string[]

Accepts the wildcard (*) character for headers matching a specified pattern.

Specifying "*" results in all headers included in the log event.

These headers are not included in the JSON payload if a field with a conflicting name exists.

Default:
[]
host_key string

An optional path that deserializes an empty string to None.

keepalive object

Configuration of HTTP server keepalive parameters.

2 nested properties
max_connection_age_jitter_factor number

A value of 0.1 means that the actual duration will be between 90% and 110% of the specified maximum duration.

Default: 0.1
min=0.0max=1.0
max_connection_age_secs integer | null

Only applies to HTTP/0.9, HTTP/1.0, and HTTP/1.1 requests.

A random jitter configured by max_connection_age_jitter_factor is added to the specified duration to spread out connection storms.

Default: 300
min=0.0max=9007199254740990.0
log_namespace boolean | null

The namespace to use for logs. This overrides the global setting.

Default: null
method const: "HEAD" | const: "GET" | const: "POST" | const: "PUT" | const: "PATCH" | const: "DELETE" | const: "OPTIONS"

HTTP method.

path string

The URL path on which log event POST requests are sent.

Default: "/"
path_key string

An optional path that deserializes an empty string to None.

query_parameters string[]

Accepts the wildcard (*) character for query parameters matching a specified pattern.

Specifying "*" results in all query parameters included in the log event.

These override any values included in the body with conflicting names.

Default:
[]
response_code integer

Specifies the HTTP response status code that will be returned on successful requests.

Default: 200
min=0.0max=65535.0
strict_path boolean

If set to true, only requests using the exact URL path specified in path are accepted. Otherwise, requests sent to a URL path that starts with the value of path are accepted.

With strict_path set to false and path set to "", the configured HTTP source accepts requests from any URL path.

Default: true
vector::sources::internal_logs::InternalLogsConfig object

Configuration for the internal_logs source.

log_namespace boolean | null

The namespace to use for logs. This overrides the global setting.

Default: null
pid_key string

An optional path that deserializes an empty string to None.

vector::sources::internal_metrics::InternalMetricsConfig object

Configuration for the internal_metrics source.

namespace string

Overrides the default namespace for the metrics emitted by the source.

Default: "vector"
scrape_interval_secs number

A span of time, in fractional seconds.

min=-9007199254740990.0max=9007199254740990.0
tags object

Tag configuration for the internal_metrics source.

Default:
{
  "host_key": null,
  "pid_key": null
}
2 nested properties

The value is the peer host's address, including the port. For example, 1.2.3.4:9000.

By default, the global log_schema.host_key option is used.

Set to "" to suppress this key.

pid_key string | null

By default, this is not set and the tag is not automatically added.

vector::sources::journald::JournaldConfig object

Configuration for the journald source.

acknowledgements object

This setting is deprecated in favor of enabling acknowledgements at the global or sink level.

Enabling or disabling acknowledgements at the source level has no effect on acknowledgement behavior.

See End-to-end Acknowledgements for more information on how event acknowledgement is handled.

1 nested properties
enabled boolean | null

Whether or not end-to-end acknowledgements are enabled for this source.

batch_size integer

This option limits the size of the batch.

Default: 16
min=0.0max=9007199254740990.0
current_boot_only boolean

Only include entries that occurred after the current boot of the system.

Default: true

By default, the global data_dir option is used. Make sure the running user has write permissions to this directory.

If this directory is specified, then Vector will attempt to create it.

Default: null
emit_cursor boolean
Default: false
exclude_matches Record<string, string[]>

If exclude_units is specified, it is merged into this list.

Default:
{}
exclude_units string[]

Unit names lacking a . have .service appended to make them a valid service unit name.

Default:
[]
extra_args string[]

If specified, it is merged to the command line arguments as-is.

Default:
[]
include_matches Record<string, string[]>

If empty or not present, all journal fields are accepted.

If include_units is specified, it is merged into this list.

Default:
{}
include_units string[]

If empty or not present, all units are accepted.

Unit names lacking a . have .service appended to make them a valid service unit name.

Default:
[]

If not set, journalctl uses the default system journal path.

Default: null
journal_namespace string | null

This value is passed to journalctl through the --namespace option. If not set, journalctl uses the default namespace.

Default: null

If not set, a search is done for the journalctl path.

Default: null
log_namespace boolean | null

The namespace to use for logs. This overrides the global setting.

Default: null
remap_priority boolean

Has no effect unless the value of the field is already an integer.

Default: false
since_now boolean

Only include entries that appended to the journal after the entries have been read.

Default: false
vector::sources::kafka::KafkaSourceConfig

Configuration for the kafka source.

vector::sources::kubernetes_logs::Config object

Configuration for the kubernetes_logs source.

auto_partial_merge boolean

Partial events are messages that were split by the Kubernetes Container Runtime log driver.

By default, the global data_dir option is used. Make sure the running user has write permissions to this directory.

If this directory is specified, then Vector will attempt to create it.

delay_deletion_ms integer

A span of time, in whole milliseconds.

min=0.0max=9007199254740990.0
exclude_paths_glob_patterns stdlib::PathBuf[]

A list of glob patterns to exclude from reading the files.

extra_field_selector string

The built-in Node filter uses self_node_name to only watch Pods located on the same Node.

extra_label_selector string
extra_namespace_label_selector string
fingerprint_lines integer

If your files share a common header that is not always a fixed size,

If the file has less than this amount of lines, it won’t be read at all.

min=0.0max=9007199254740990.0
glob_minimum_cooldown_ms integer

A span of time, in whole milliseconds.

min=0.0max=9007199254740990.0
ignore_older_secs integer | null

Ignore files with a data modification date older than the specified number of seconds.

Default: null
min=0.0max=9007199254740990.0
include_paths_glob_patterns stdlib::PathBuf[]

A list of glob patterns to include while reading the files.

This is useful to compute the latency between important event processing stages. For example, the time delta between when a log line was written and when it was processed by the kubernetes_logs source.

internal_metrics object

Configuration of internal metrics for file-based components.

1 nested properties
include_file_tag boolean

This is useful for distinguishing between different files while monitoring. However, the tag's cardinality is unbounded.

Default: false

If not set, a connection to Kubernetes is made using the in-cluster configuration.

log_namespace boolean | null

The namespace to use for logs. This overrides the global setting.

Default: null
max_line_bytes integer

This protects against malformed lines or tailing incorrect files.

min=0.0max=9007199254740990.0
max_read_bytes integer

This allows distributing the reads more or less evenly across the files.

min=0.0max=9007199254740990.0
namespace_annotation_fields object

Configuration for how the events are enriched with Namespace metadata.

Default:
{
  "namespace_labels": ".kubernetes.namespace_labels"
}
1 nested properties
namespace_labels string

An optional path that deserializes an empty string to None.

node_annotation_fields object

Configuration for how the events are enriched with Node metadata.

Default:
{
  "node_labels": ".kubernetes.node_labels"
}
1 nested properties
node_labels string

An optional path that deserializes an empty string to None.

oldest_first boolean

Instead of balancing read capacity fairly across all watched files, prioritize draining the oldest files before moving on to read data from more recent files.

Default: true
pod_annotation_fields object

Configuration for how the events are enriched with Pod metadata.

Default:
{
  "pod_name": ".kubernetes.pod_name",
  "pod_namespace": ".kubernetes.pod_namespace",
  "pod_uid": ".kubernetes.pod_uid",
  "pod_ip": ".kubernetes.pod_ip",
  "pod_ips": ".kubernetes.pod_ips",
  "pod_labels": ".kubernetes.pod_labels",
  "pod_annotations": ".kubernetes.pod_annotations",
  "pod_node_name": ".kubernetes.pod_node_name",
  "pod_owner": ".kubernetes.pod_owner",
  "container_name": ".kubernetes.container_name",
  "container_id": ".kubernetes.container_id",
  "container_image": ".kubernetes.container_image",
  "container_image_id": ".kubernetes.container_image_id"
}
13 nested properties
container_id string

An optional path that deserializes an empty string to None.

container_image string

An optional path that deserializes an empty string to None.

container_image_id string

An optional path that deserializes an empty string to None.

container_name string

An optional path that deserializes an empty string to None.

pod_annotations string

An optional path that deserializes an empty string to None.

pod_ip string

An optional path that deserializes an empty string to None.

pod_ips string

An optional path that deserializes an empty string to None.

pod_labels string

An optional path that deserializes an empty string to None.

pod_name string

An optional path that deserializes an empty string to None.

pod_namespace string

An optional path that deserializes an empty string to None.

pod_node_name string

An optional path that deserializes an empty string to None.

pod_owner string

An optional path that deserializes an empty string to None.

pod_uid string

An optional path that deserializes an empty string to None.

read_from const: "beginning" | const: "end"

File position to use when reading a new file.

rotate_wait_secs integer

A span of time, in whole seconds.

min=0.0max=9007199254740990.0
self_node_name string

Configured to use an environment variable by default, to be evaluated to a value provided by Kubernetes at Pod creation.

The default time zone for timestamps without an explicit zone.

use_apiserver_cache boolean

Determines if requests to the kube-apiserver can be served by a cache.

vector::sources::logstash::LogstashConfig object

Configuration for the logstash source.

address string required

If a socket address is used, it must include a port.

acknowledgements object

This setting is deprecated in favor of enabling acknowledgements at the global or sink level.

Enabling or disabling acknowledgements at the source level has no effect on acknowledgement behavior.

See End-to-end Acknowledgements for more information on how event acknowledgement is handled.

1 nested properties
enabled boolean | null

Whether or not end-to-end acknowledgements are enabled for this source.

connection_limit integer | null

The maximum number of TCP connections that are allowed at any given time.

min=0.0max=4294967295.0
log_namespace boolean | null

The namespace to use for logs. This overrides the global setting.

Default: null
receive_buffer_bytes integer | null

The size of the receive buffer used for each connection.

min=0.0max=9007199254740990.0
vector::sources::mongodb_metrics::MongoDbMetricsConfig object

Configuration for the mongodb_metrics source.

endpoints string[] required

Each endpoint must be in the Connection String URI Format.

namespace string

If set to an empty string, no namespace is added to the metrics.

By default, mongodb is used.

Default: "mongodb"
scrape_interval_secs integer

A span of time, in whole seconds.

min=0.0max=9007199254740990.0
vector::sources::nats::NatsSourceConfig object

Configuration for the nats source.

connection_name string required
subject string required
url string required

The URL takes the form of nats://server:port. If the port is not specified it defaults to 4222.

decoding object | object | object

Configures how events are decoded from raw bytes.

framing object

Framing handles how events are separated when encoded in a raw byte form, where each event is a frame that must be prefixed, or delimited, in a way that marks where an event begins and ends within the byte stream.

log_namespace boolean | null

The namespace to use for logs. This overrides the global setting.

Default: null
queue string | null

The NATS queue group to join.

subject_key_field string

An optional path that deserializes an empty string to None.

subscriber_capacity integer

This value determines how many messages the NATS subscriber buffers before incoming messages are dropped.

See the async_nats documentation for more information.

Default: 65536
min=0.0max=9007199254740990.0
vector::sources::nginx_metrics::NginxMetricsConfig object

Configuration for the nginx_metrics source.

endpoints string[] required

Each endpoint must be a valid HTTP/HTTPS URI pointing to an NGINX instance that has the ngx_http_stub_status_module module enabled.

namespace string

If set to an empty string, no namespace is added to the metrics.

By default, nginx is used.

Default: "nginx"
scrape_interval_secs integer

A span of time, in whole seconds.

min=0.0max=9007199254740990.0
vector::sources::opentelemetry::OpentelemetryConfig object

Configuration for the opentelemetry source.

grpc object required

Configuration for the opentelemetry gRPC server.

2 nested properties
address string required

An internet socket address, either IPv4 or IPv6.

http object required

Configuration for the opentelemetry HTTP server.

4 nested properties
address string required

An internet socket address, either IPv4 or IPv6.

headers string[]

Accepts the wildcard (*) character for headers matching a specified pattern.

Specifying "*" results in all headers included in the log event.

These headers are not included in the JSON payload if a field with a conflicting name exists.

Default:
[]
keepalive object

Configuration of HTTP server keepalive parameters.

2 nested properties
max_connection_age_jitter_factor number

A value of 0.1 means that the actual duration will be between 90% and 110% of the specified maximum duration.

Default: 0.1
min=0.0max=1.0
max_connection_age_secs integer | null

Only applies to HTTP/0.9, HTTP/1.0, and HTTP/1.1 requests.

A random jitter configured by max_connection_age_jitter_factor is added to the specified duration to spread out connection storms.

Default: 300
min=0.0max=9007199254740990.0
acknowledgements object

This setting is deprecated in favor of enabling acknowledgements at the global or sink level.

Enabling or disabling acknowledgements at the source level has no effect on acknowledgement behavior.

See End-to-end Acknowledgements for more information on how event acknowledgement is handled.

1 nested properties
enabled boolean | null

Whether or not end-to-end acknowledgements are enabled for this source.

log_namespace boolean | null

The namespace to use for logs. This overrides the global setting.

Default: null
vector::sources::postgresql_metrics::PostgresqlMetricsConfig object

Configuration for the postgresql_metrics source.

endpoints string[] required

Each endpoint must be in the Connection URI format.

exclude_databases array | null

Specifying "" includes metrics where datname is NULL.

This can be used in conjunction with include_databases.

include_databases array | null

If not set, metrics are collected from all databases. Specifying "" includes metrics where datname is NULL.

This can be used in conjunction with exclude_databases.

namespace string

Overrides the default namespace for the metrics emitted by the source.

Default: "postgresql"
scrape_interval_secs integer

A span of time, in whole seconds.

min=0.0max=9007199254740990.0
tls null | object

Configuration of TLS when connecting to PostgreSQL.

vector::sources::prometheus::pushgateway::PrometheusPushgatewayConfig object

Configuration for the prometheus_pushgateway source.

address string required

An internet socket address, either IPv4 or IPv6.

acknowledgements object

This setting is deprecated in favor of enabling acknowledgements at the global or sink level.

Enabling or disabling acknowledgements at the source level has no effect on acknowledgement behavior.

See End-to-end Acknowledgements for more information on how event acknowledgement is handled.

1 nested properties
enabled boolean | null

Whether or not end-to-end acknowledgements are enabled for this source.

aggregate_metrics boolean

Only applies to counters and histograms as gauges and summaries can't be meaningfully aggregated.

Default: false
keepalive object

Configuration of HTTP server keepalive parameters.

2 nested properties
max_connection_age_jitter_factor number

A value of 0.1 means that the actual duration will be between 90% and 110% of the specified maximum duration.

Default: 0.1
min=0.0max=1.0
max_connection_age_secs integer | null

Only applies to HTTP/0.9, HTTP/1.0, and HTTP/1.1 requests.

A random jitter configured by max_connection_age_jitter_factor is added to the specified duration to spread out connection storms.

Default: 300
min=0.0max=9007199254740990.0
vector::sources::prometheus::remote_write::PrometheusRemoteWriteConfig object

Configuration for the prometheus_remote_write source.

address string required

An internet socket address, either IPv4 or IPv6.

acknowledgements object

This setting is deprecated in favor of enabling acknowledgements at the global or sink level.

Enabling or disabling acknowledgements at the source level has no effect on acknowledgement behavior.

See End-to-end Acknowledgements for more information on how event acknowledgement is handled.

1 nested properties
enabled boolean | null

Whether or not end-to-end acknowledgements are enabled for this source.

keepalive object

Configuration of HTTP server keepalive parameters.

2 nested properties
max_connection_age_jitter_factor number

A value of 0.1 means that the actual duration will be between 90% and 110% of the specified maximum duration.

Default: 0.1
min=0.0max=1.0
max_connection_age_secs integer | null

Only applies to HTTP/0.9, HTTP/1.0, and HTTP/1.1 requests.

A random jitter configured by max_connection_age_jitter_factor is added to the specified duration to spread out connection storms.

Default: 300
min=0.0max=9007199254740990.0
vector::sources::prometheus::scrape::PrometheusScrapeConfig object

Configuration for the prometheus_scrape source.

endpoints string[] required

Endpoints to scrape metrics from.

endpoint_tag string | null

The tag value is the endpoint of the scraped instance.

honor_labels boolean

If true, the new tag is not added if the scraped metric has the tag already. If false, the conflicting tag is renamed by prepending exported_ to the original name.

This matches Prometheus’ honor_labels configuration.

Default: false
instance_tag string | null

The tag value is the host and port of the scraped instance.

query Record<string, string[]>

One or more values for the same parameter key can be provided. The parameters provided in this option are appended to any parameters manually provided in the endpoints option. This option is especially useful when scraping the /federate endpoint.

Default:
{}
scrape_interval_secs integer

A span of time, in whole seconds.

min=0.0max=9007199254740990.0
scrape_timeout_secs number

A span of time, in fractional seconds.

min=-9007199254740990.0max=9007199254740990.0
vector::sources::pulsar::PulsarSourceConfig object

Configuration for the pulsar source.

endpoint string required

The endpoint to which the Pulsar client should connect to.

topics string[] required

The Pulsar topic names to read events from.

acknowledgements object

This setting is deprecated in favor of enabling acknowledgements at the global or sink level.

Enabling or disabling acknowledgements at the source level has no effect on acknowledgement behavior.

See End-to-end Acknowledgements for more information on how event acknowledgement is handled.

1 nested properties
enabled boolean | null

Whether or not end-to-end acknowledgements are enabled for this source.

auth null | object | object

Authentication configuration.

batch_size integer | null

Max count of messages in a batch.

min=0.0max=4294967295.0
consumer_name string | null

The Pulsar consumer name.

dead_letter_queue_policy null | object

Dead Letter Queue policy configuration.

decoding object | object | object

Configures how events are decoded from raw bytes.

framing object

Framing handles how events are separated when encoded in a raw byte form, where each event is a frame that must be prefixed, or delimited, in a way that marks where an event begins and ends within the byte stream.

log_namespace boolean | null

The namespace to use for logs. This overrides the global setting.

Default: null
priority_level integer | null

The broker follows descending priorities. For example, 0=max-priority, 1, 2,...

In Shared subscription type, the broker first dispatches messages to the max priority level consumers if they have permits. Otherwise, the broker considers next priority level consumers.

min=-2147483648.0max=2147483647.0
subscription_name string | null

The Pulsar subscription name.

vector::sources::redis::RedisSourceConfig object

Configuration for the redis source.

key string required

The Redis key to read messages from.

url string required

The URL must take the form of protocol://server:port/db where the protocol can either be redis or rediss for connections secured using TLS.

data_type const: "list" | const: "channel"

The Redis data type (list or channel) to use.

Default: "list"
decoding object | object | object

Configures how events are decoded from raw bytes.

framing object

Framing handles how events are separated when encoded in a raw byte form, where each event is a frame that must be prefixed, or delimited, in a way that marks where an event begins and ends within the byte stream.

list null | object

Options for the Redis list data type.

log_namespace boolean | null

The namespace to use for logs. This overrides the global setting.

Default: null

The value is the Redis key that the event was read from.

By default, this is not set and the field is not automatically added.

vector::sources::socket::SocketConfig

Configuration for the socket source.

vector::sources::splunk_hec::SplunkConfig object

Configuration for the splunk_hec source.

acknowledgements object

Acknowledgement configuration for the splunk_hec source.

Default:
{
  "enabled": null,
  "max_pending_acks": 10000000,
  "max_number_of_ack_channels": 1000000,
  "max_pending_acks_per_channel": 1000000,
  "ack_idle_cleanup": false,
  "max_idle_time": 300
}
6 nested properties
ack_idle_cleanup boolean

A channel is idling if it is not used for sending data or querying acknowledgement statuses.

enabled boolean | null

Enables end-to-end acknowledgements.

max_idle_time integer

Channels can potentially idle for longer than this setting but clients should not rely on such behavior.

Minimum of 1.

min=1.0max=9007199254740990.0
max_number_of_ack_channels integer

Minimum of 1.

min=1.0max=9007199254740990.0
max_pending_acks integer

Equivalent to the max_number_of_acked_requests_pending_query Splunk HEC setting.

Minimum of 1.

min=1.0max=9007199254740990.0
max_pending_acks_per_channel integer

Equivalent to the max_number_of_acked_requests_pending_query_per_ack_channel Splunk HEC setting.

Minimum of 1.

min=1.0max=9007199254740990.0
address string

An internet socket address, either IPv4 or IPv6.

keepalive object

Configuration of HTTP server keepalive parameters.

2 nested properties
max_connection_age_jitter_factor number

A value of 0.1 means that the actual duration will be between 90% and 110% of the specified maximum duration.

Default: 0.1
min=0.0max=1.0
max_connection_age_secs integer | null

Only applies to HTTP/0.9, HTTP/1.0, and HTTP/1.1 requests.

A random jitter configured by max_connection_age_jitter_factor is added to the specified duration to spread out connection storms.

Default: 300
min=0.0max=9007199254740990.0
log_namespace boolean | null

The namespace to use for logs. This overrides the global settings.

Default: null
store_hec_token boolean

If set to true, when incoming requests contain a Splunk HEC token, the token used is kept in the event metadata and preferentially used if the event is sent to a Splunk HEC sink.

If supplied, incoming requests must supply this token in the Authorization header, just as a client would if it was communicating with the Splunk HEC endpoint directly.

If not supplied, the Authorization header is ignored and requests are not authenticated.

valid_tokens array | null

If supplied, incoming requests must supply one of these tokens in the Authorization header, just as a client would if it was communicating with the Splunk HEC endpoint directly.

If not supplied, the Authorization header is ignored and requests are not authenticated.

vector::sources::static_metrics::StaticMetricsConfig object

Configuration for the static_metrics source.

interval_secs number

A span of time, in fractional seconds.

min=-9007199254740990.0max=9007199254740990.0
metrics object[]

Tag configuration for the internal_metrics source.

Default:
[]
namespace string

Overrides the default namespace for the metrics emitted by the source.

Default: "static"
vector::sources::statsd::StatsdConfig

Configuration for the statsd source.

vector::sources::syslog::SyslogConfig

Configuration for the syslog source.

vector::sources::util::http::method::HttpMethod const: "HEAD" | const: "GET" | const: "POST" | const: "PUT" | const: "PATCH" | const: "DELETE" | const: "OPTIONS"

HTTP method.

vector::sources::util::net::SocketListenAddr string

If a socket address is used, it must include a port.

vector::sources::vector::VectorConfig object

Configuration for the vector source.

address string required

An internet socket address, either IPv4 or IPv6.

acknowledgements object

This setting is deprecated in favor of enabling acknowledgements at the global or sink level.

Enabling or disabling acknowledgements at the source level has no effect on acknowledgement behavior.

See End-to-end Acknowledgements for more information on how event acknowledgement is handled.

1 nested properties
enabled boolean | null

Whether or not end-to-end acknowledgements are enabled for this source.

log_namespace boolean | null

The namespace to use for logs. This overrides the global setting.

Default: null
version null | const: "2"

Version of the configuration.

vector::template::Template string

In many cases, components can be configured so that part of the component's functionality can be customized on a per-event basis. For example, you have a sink that writes events to a file and you want to specify which file an event should go to by using an event field as part of the input to the filename used.

By using Template, users can specify either fixed strings or templated strings. Templated strings use a common syntax to refer to fields in an event that is used as the input data when rendering the template. An example of a fixed string is my-file.log. An example of a template string is my-file-{{key}}.log, where {{key}} is the key's value when the template is rendered into a string.

vector::transforms::Transforms

Configurable transforms in Vector.

vector::transforms::aggregate::AggregateConfig object

Configuration for the aggregate transform.

interval_ms integer

During this time frame, metrics (beta) with the same series data (name, namespace, tags, and so on) are aggregated.

Default: 10000
min=0.0max=9007199254740990.0
mode const: "Auto" | const: "Sum" | const: "Latest" | const: "Count" | const: "Diff" | const: "Max" | const: "Min" | const: "Mean" | const: "Stdev"

Some of the functions may only function on incremental and some only on absolute metrics.

Default: "Auto"
vector::transforms::aws_ec2_metadata::Ec2Metadata object

Configuration for the aws_ec2_metadata transform.

endpoint string

Overrides the default EC2 metadata endpoint.

Default: "http://169.254.169.254"
fields string[]

A list of metadata fields to include in each transformed event.

Default:
[
  "ami-id",
  "availability-zone",
  "instance-id",
  "instance-type",
  "local-hostname",
  "local-ipv4",
  "public-hostname",
  "public-ipv4",
  "region",
  "subnet-id",
  "vpc-id",
  "role-name"
]

Sets a prefix for all event fields added by the transform.

proxy object

Configure to proxy traffic through an HTTP(S) proxy when making external requests.

Similar to common proxy configuration convention, you can set different proxies to use based on the type of traffic being proxied. You can also set specific hosts that should not be proxied.

4 nested properties
enabled boolean

Enables proxying support.

Default: true
http string | null

Must be a valid URI string.

Default: null
format=uri
https string | null

Must be a valid URI string.

Default: null
format=uri
no_proxy string[]

Multiple patterns are allowed:

PatternExample match
Domain namesexample.com matches requests to example.com
Wildcard domains.example.com matches requests to example.com and its subdomains
IP addresses127.0.0.1 matches requests to 127.0.0.1
CIDR blocks192.168.0.0/16 matches requests to any IP addresses in this range
Splat* matches all hosts
Default:
[]
refresh_interval_secs integer

A span of time, in whole seconds.

min=0.0max=9007199254740990.0
refresh_timeout_secs integer

A span of time, in whole seconds.

min=0.0max=9007199254740990.0
required boolean

Requires the transform to be able to successfully query the EC2 metadata before starting to process the data.

Default: true
tags string[]

A list of instance tags to include in each transformed event.

Default:
[]
vector::transforms::dedupe::config::DedupeConfig object

Configuration for the dedupe transform.

cache object

Caching configuration for deduplication.

Default:
{
  "num_events": 5000
}
1 nested properties
num_events integer required

Number of events to cache and use for comparing incoming events to previously seen events.

min=1.0max=9007199254740990.0
fields null | object | object

When no field matching configuration is specified, events are matched using the timestamp, host, and message fields from an event. The specific field names used are those set in the global log schema configuration.

Default: null
vector::transforms::exclusive_route::config::ExclusiveRouteConfig object

Configuration for the route transform.

routes object[] required

An array of named routes. The route names are expected to be unique.

vector::transforms::filter::FilterConfig object

Configuration for the filter transform.

condition string | object | object | object required

Many methods exist for matching events, such as using a VRL expression, a Datadog Search query string, or hard-coded matchers like "must be a metric" or "fields A, B, and C must match these constraints".

As VRL is the most common way to apply conditions to events, this type provides a shortcut to define VRL expressions directly in the configuration by passing the VRL expression as a string:

condition = '.message == "hooray"'

When other condition types are required, they can be specified with an enum-style notation:

condition.type = 'datadog_search'
condition.source = 'NOT "foo"'
vector::transforms::log_to_metric::LogToMetricConfig object

Configuration for the log_to_metric transform.

metrics array required

A list of metrics to generate.

all_metrics boolean | null

This is an example JSON representation of a counter with the following properties:

  • counter: An object with a single property value representing the counter value, in this case, 10.0).
  • kind: A string indicating the kind of counter, in this case, "incremental".
  • name: A string representing the name of the counter, here set to "test.transform.counter".
  • tags: An object containing additional tags such as "env" and "host".

Objects that can be processed include counter, histogram, gauge, set and summary.

vector::transforms::lua::LuaConfig

Configuration for the lua transform.

vector::transforms::metric_to_log::MetricToLogConfig object

Configuration for the metric_to_log transform.

host_tag string | null

If present, the value of the tag is set on the generated log event in the host field, where the field key uses the global host_key option.

log_namespace boolean | null

The namespace to use for logs. This overrides the global setting.

Default: null
metric_tag_values const: "single" | const: "full"

The user configuration to choose the metric tag strategy.

This overrides the global timezone option. The time zone name may be any name in the TZ database or local to indicate system local time.

vector::transforms::reduce::config::ReduceConfig object

Configuration for the reduce transform.

end_every_period_ms integer | null

If supplied, every time this interval elapses for a given grouping, the reduced value for that grouping is flushed. Checked every flush_period_ms.

Default: null
min=0.0max=9007199254740990.0

If this condition resolves to true for an event, the current transaction is immediately flushed with this event.

expire_after_ms integer

A span of time, in whole milliseconds.

min=0.0max=9007199254740990.0
flush_period_ms integer

A span of time, in whole milliseconds.

min=0.0max=9007199254740990.0
group_by string[]

Each group with matching values for the specified keys is reduced independently, allowing you to keep independent event streams separate. When no fields are specified, all events are combined in a single group.

For example, if group_by = ["host", "region"], then all incoming events that have the same host and region are grouped together before being reduced.

Default:
[]
max_events integer | null

The maximum number of events to group together.

min=1.0max=9007199254740990.0
merge_strategies Record<string, const: "discard" | const: "retain" | const: "sum" | const: "max" | const: "min" | const: "array" | const: "concat" | const: "concat_newline" | const: "concat_raw" | const: "shortest_array" | const: "longest_array" | const: "flat_unique">

For each field specified, the given strategy is used for combining events rather than the default behavior.

The default behavior is as follows:

  • The first value of a string field is kept and subsequent values are discarded.
  • For timestamp fields the first is kept and a new field [field-name]_end is added with the last received timestamp value.
  • Numeric values are summed.
  • For nested paths, the field value is retrieved and then reduced using the default strategies mentioned above (unless explicitly specified otherwise).
Default:
{}

If this condition resolves to true for an event, the previous transaction is flushed (without this event) and a new transaction is started.

vector::transforms::remap::RemapConfig object

Configuration for the remap transform.

drop_on_abort boolean

If a VRL program is manually aborted (using abort) when processing an event, this option controls whether the original, unmodified event is sent downstream without any modifications or if it is dropped.

Additionally, dropped events can potentially be diverted to a specially-named output for further logging and analysis by setting reroute_dropped.

Default: true
drop_on_error boolean

Normally, if a VRL program encounters an error when processing an event, the original, unmodified event is sent downstream. In some cases, you may not want to send the event any further, such as if certain transformation or enrichment is strictly required. Setting drop_on_error to true allows you to ensure these events do not get processed any further.

Additionally, dropped events can potentially be diverted to a specially named output for further logging and analysis by setting reroute_dropped.

Default: false

If a relative path is provided, its root is the current working directory.

Required if source is missing.

files array | null

If a relative path is provided, its root is the current working directory.

Required if source or file are missing.

metric_tag_values const: "single" | const: "full"

The user configuration to choose the metric tag strategy.

reroute_dropped boolean

When using drop_on_error or drop_on_abort, events that are "dropped" are processed no further. In some cases, it may be desirable to keep the events around for further analysis, debugging, or retrying.

In these cases, reroute_dropped can be set to true which forwards the original event to a specially-named output, dropped. The original event is annotated with additional fields describing why the event was dropped.

Default: false
runtime string

The runtime to use for executing VRL code.

Default: "ast"
source string | null

Required if file is missing.

This overrides the global timezone option. The time zone name may be any name in the TZ database, or local to indicate system local time.

Default: null
vector::transforms::route::RouteConfig object

Configuration for the route transform.

reroute_unmatched boolean

Normally, if an event doesn't match any defined route, it is sent to the <transform_name>._unmatched output for further processing. In some cases, you may want to simply discard unmatched events and not process them any further.

In these cases, reroute_unmatched can be set to false to disable the <transform_name>._unmatched output and instead silently discard any unmatched events.

Default: true
route Record<string, string | object | object | object>

The following identifiers are reserved output names and thus cannot be used as route IDs:

  • _unmatched
  • _default

Each route can then be referenced as an input by other components with the name <transform_name>.<route_id>. If an event doesn’t match any route, and if reroute_unmatched is set to true (the default), it is sent to the <transform_name>._unmatched output. Otherwise, the unmatched event is instead silently discarded.

vector::transforms::sample::config::SampleConfig object

Configuration for the sample transform.

rate integer required

For example, rate = 1500 means 1 out of every 1500 events are forwarded and the rest are dropped.

min=0.0max=9007199254740990.0

A logical condition used to exclude events from sampling.

If left unspecified, or if the event doesn't have group_by, then the event is not sampled separately.

key_field string | null

Each unique value for the key creates a bucket of related events to be sampled together and the rate is applied to the buckets themselves to sample 1/N buckets. The overall rate of sampling may differ from the configured one if values in the field are not uniformly distributed. If left unspecified, or if the event doesn’t have key_field, then the event is sampled independently.

This can be useful to, for example, ensure that all logs for a given transaction are sampled together, but that overall 1/N transactions are sampled.

sample_rate_key string

An optional path that deserializes an empty string to None.

vector::transforms::tag_cardinality_limit::config::TagCardinalityLimitConfig

Configuration for the tag_cardinality_limit transform.

vector::transforms::throttle::ThrottleConfig object

Configuration for the throttle transform.

threshold integer required

Each unique key has its own threshold.

min=0.0max=4294967295.0
window_secs number required

A span of time, in fractional seconds.

min=-9007199254740990.0max=9007199254740990.0

A logical condition used to exclude events from sampling.

internal_metrics object

Configuration of internal metrics for the Throttle transform.

Default:
{
  "emit_events_discarded_per_key": false
}
1 nested properties
emit_events_discarded_per_key boolean

If true, the counter will be incremented for each discarded event, including the key value associated with the discarded event. If false, the counter will not be emitted. Instead, the number of discarded events can be seen through the component_discarded_events_total internal metric.

Note that this defaults to false because the key tag has potentially unbounded cardinality. Only set this to true if you know that the number of unique keys is bounded.

Default: false

If left unspecified, or if the event doesn't have key_field, then the event is not rate limited separately.

vector_common::id::Inputs<alloc::string::String> string[]

Wildcards (*) are supported.

See configuration for more info.

vector_common::sensitive_string::SensitiveString string

Wrapper for sensitive strings containing credentials

vector_core::config::AcknowledgementsConfig object

See End-to-end Acknowledgements for more information on how event acknowledgement is handled.

enabled boolean | null

When enabled for a sink, any source connected to that sink, where the source supports end-to-end acknowledgements as well, waits for events to be acknowledged by all connected sinks before acknowledging them at the source.

Enabling or disabling acknowledgements at the sink level takes precedence over any global acknowledgements configuration.

vector_core::config::SourceAcknowledgementsConfig object

This setting is deprecated in favor of enabling acknowledgements at the global or sink level.

Enabling or disabling acknowledgements at the source level has no effect on acknowledgement behavior.

See End-to-end Acknowledgements for more information on how event acknowledgement is handled.

enabled boolean | null

Whether or not end-to-end acknowledgements are enabled for this source.

vector_core::config::proxy::ProxyConfig object

Configure to proxy traffic through an HTTP(S) proxy when making external requests.

Similar to common proxy configuration convention, you can set different proxies to use based on the type of traffic being proxied. You can also set specific hosts that should not be proxied.

enabled boolean

Enables proxying support.

Default: true
http string | null

Must be a valid URI string.

Default: null
format=uri
https string | null

Must be a valid URI string.

Default: null
format=uri
no_proxy string[]

Multiple patterns are allowed:

PatternExample match
Domain namesexample.com matches requests to example.com
Wildcard domains.example.com matches requests to example.com and its subdomains
IP addresses127.0.0.1 matches requests to 127.0.0.1
CIDR blocks192.168.0.0/16 matches requests to any IP addresses in this range
Splat* matches all hosts
Default:
[]
vector_core::event::metric::MetricKind const: "incremental" | const: "absolute"

Metrics can be either absolute or incremental. Absolute metrics represent a sort of "last write wins" scenario, where the latest absolute value seen is meant to be the actual metric value. In contrast, and perhaps intuitively, incremental metrics are meant to be additive, such that we don't know what total value of the metric is, but we know that we'll be adding or subtracting the given value from it.

Generally speaking, most metrics storage systems deal with incremental updates. A notable exception is Prometheus, which deals with, and expects, absolute values from clients.

vector_core::event::metric::value::MetricValue object | object | object | object | object | object | object

Metric value. Container for the actual value of a metric.

vector_core::tls::settings::TlsConfig object

TLS configuration.

alpn_protocols array | null

Declare the supported ALPN protocols, which are used during negotiation with peer. They are prioritized in the order that they are defined.

The certificate must be in the DER or PEM (X.509) format. Additionally, the certificate can be provided as an inline string in PEM format.

The certificate must be in DER, PEM (X.509), or PKCS#12 format. Additionally, the certificate can be provided as an inline string in PEM format.

If this is set, and is not a PKCS#12 archive, key_file must also be set.

The key must be in DER or PEM (PKCS#8) format. Additionally, the key can be provided as an inline string in PEM format.

key_pass string | null

This has no effect unless key_file is set.

server_name string | null

Only relevant for outgoing connections.

verify_certificate boolean | null

If enabled, certificates must not be expired and must be issued by a trusted issuer. This verification operates in a hierarchical manner, checking that the leaf certificate (the certificate presented by the client/server) is not only valid, but that the issuer of that certificate is also valid, and so on until the verification process reaches a root certificate.

Do NOT set this to false unless you understand the risks of not verifying the validity of certificates.

verify_hostname boolean | null

If enabled, the hostname used to connect to the remote host must be present in the TLS certificate presented by the remote host, either as the Common Name or as an entry in the Subject Alternative Name extension.

Only relevant for outgoing connections.

Do NOT set this to false unless you understand the risks of not verifying the remote hostname.

vector_core::tls::settings::TlsEnableableConfig

Configures the TLS options for incoming/outgoing connections.

vector_lookup::lookup_v2::ConfigTargetPath string

A wrapper around OwnedTargetPath that allows it to be used in Vector config with prefix default to PathPrefix::Event

vector_lookup::lookup_v2::ConfigValuePath string

A wrapper around OwnedValuePath that allows it to be used in Vector config. This requires a valid path to be used. If you want to allow optional paths, use [optional_path::OptionalValuePath].

vector_lookup::lookup_v2::optional_path::OptionalTargetPath string

An optional path that deserializes an empty string to None.

vector_lookup::lookup_v2::optional_path::OptionalValuePath string

An optional path that deserializes an empty string to None.