gRPC exposes a variety of settings. Several of these settings are available for configuration within individual receivers or exporters. In general, none of these settings should need to be adjusted.
Exporters leverage client configuration.
Note that client configuration supports TLS configuration, the
configuration parameters are also defined under tls
like server
configuration. For more information, see configtls
README.
balancer_name
: Default before v0.103.0 ispick_first
, default for v0.103.0 isround_robin
. See issue. To restore the previous behavior, setbalancer_name
topick_first
.compression
: Compression type to use amonggzip
,snappy
,zstd
, andnone
.endpoint
: Valid value syntax available heretls
headers
: name/value pairs added to the requestkeepalive
permit_without_stream
time
timeout
read_buffer_size
write_buffer_size
auth
Please note that per_rpc_auth
which allows the credentials to send for every RPC is now moved to become an extension. Note that this feature isn't about sending the headers only during the initial connection as an authorization
header under the headers
would do: this is sent for every RPC performed during an established connection.
Example:
exporters:
otlp:
endpoint: otelcol2:55690
auth:
authenticator: some-authenticator-extension
tls:
ca_file: ca.pem
cert_file: cert.pem
key_file: key.pem
headers:
test1: "value1"
"test 2": "value 2"
configgrpc_benchmark_test.go contains benchmarks comparing the supported compression algorithms. It performs compression using gzip
, zstd
, and snappy
compression on small, medium, and large sized log, trace, and metric payloads. Each test case outputs the uncompressed payload size, the compressed payload size, and the average nanoseconds spent on compression.
The following table summarizes the results, including some additional columns computed from the raw data. The benchmarks were performed on an AWS m5.large EC2 instance with an Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz.
Request | Compressor | Raw Bytes | Compressed bytes | Compression ratio | Ns / op | Mb compressed / second | Mb saved / second |
---|---|---|---|---|---|---|---|
lg_log_request | gzip | 5150 | 262 | 19.66 | 49231 | 104.61 | 99.29 |
lg_metric_request | gzip | 6800 | 201 | 33.83 | 51816 | 131.23 | 127.35 |
lg_trace_request | gzip | 9200 | 270 | 34.07 | 65174 | 141.16 | 137.02 |
md_log_request | gzip | 363 | 268 | 1.35 | 37609 | 9.65 | 2.53 |
md_metric_request | gzip | 320 | 145 | 2.21 | 30141 | 10.62 | 5.81 |
md_trace_request | gzip | 451 | 288 | 1.57 | 38270 | 11.78 | 4.26 |
sm_log_request | gzip | 166 | 168 | 0.99 | 30511 | 5.44 | -0.07 |
sm_metric_request | gzip | 185 | 142 | 1.30 | 29055 | 6.37 | 1.48 |
sm_trace_request | gzip | 233 | 205 | 1.14 | 33466 | 6.96 | 0.84 |
lg_log_request | snappy | 5150 | 475 | 10.84 | 1915 | 2,689.30 | 2,441.25 |
lg_metric_request | snappy | 6800 | 466 | 14.59 | 2266 | 3,000.88 | 2,795.23 |
lg_trace_request | snappy | 9200 | 644 | 14.29 | 3281 | 2,804.02 | 2,607.74 |
md_log_request | snappy | 363 | 300 | 1.21 | 770.0 | 471.43 | 81.82 |
md_metric_request | snappy | 320 | 162 | 1.98 | 588.6 | 543.66 | 268.43 |
md_trace_request | snappy | 451 | 330 | 1.37 | 907.7 | 496.86 | 133.30 |
sm_log_request | snappy | 166 | 184 | 0.90 | 551.8 | 300.83 | -32.62 |
sm_metric_request | snappy | 185 | 154 | 1.20 | 526.3 | 351.51 | 58.90 |
sm_trace_request | snappy | 233 | 251 | 0.93 | 682.1 | 341.59 | -26.39 |
lg_log_request | zstd | 5150 | 223 | 23.09 | 17998 | 286.14 | 273.75 |
lg_metric_request | zstd | 6800 | 144 | 47.22 | 14289 | 475.89 | 465.81 |
lg_trace_request | zstd | 9200 | 208 | 44.23 | 17160 | 536.13 | 524.01 |
md_log_request | zstd | 363 | 261 | 1.39 | 11216 | 32.36 | 9.09 |
md_metric_request | zstd | 320 | 145 | 2.21 | 9318 | 34.34 | 18.78 |
md_trace_request | zstd | 451 | 301 | 1.50 | 12583 | 35.84 | 11.92 |
sm_log_request | zstd | 166 | 165 | 1.01 | 12482 | 13.30 | 0.08 |
sm_metric_request | zstd | 185 | 139 | 1.33 | 8824 | 20.97 | 5.21 |
sm_trace_request | zstd | 233 | 203 | 1.15 | 10134 | 22.99 | 2.96 |
Compression ratios will vary in practice as they are highly dependent on the data's information entropy. Compression rates are dependent on the speed of the CPU, and the size of payloads being compressed: smaller payloads compress at slower rates relative to larger payloads, which are able to amortize fixed computation costs over more bytes.
gzip
is the only required compression algorithm required for OTLP servers, and is a natural first choice. It is not as fast as snappy
, but achieves better compression ratios and has reasonable performance. If your collector is CPU bound and your OTLP server supports it, you may benefit from using snappy
compression. If your collector is CPU bound and has a very fast network link, you may benefit from disabling compression, which is the default.
Receivers leverage server configuration.
Note that transport configuration can also be configured. For more information, see confignet README.
keepalive
enforcement_policy
min_time
permit_without_stream
server_parameters
max_connection_age
max_connection_age_grace
max_connection_idle
time
timeout
max_concurrent_streams
max_recv_msg_size_mib
read_buffer_size
tls
write_buffer_size
auth