This package is a reference implementation of the OTLP Arrow protocol specified in this OTEP. All OTLP entities are covered (metrics, logs, and traces) as well as all sub-elements such as events, links, gauge, sum, summary, histograms, ... The overall goal is to optimize the compression ratio for telemetry data transmission as well as the end-to-end performance between telemetry data producers and receivers.
This package is still experimental and subject to change. It is currently used by an experimental OTLP/Arrow gRPC exporter and receiver.
Other important links:
- OTEP/Specification describing the motivations, the protocol, the schemas, the benchmark results and the different phases of this project. The OTEP is still pending, unmerged.
- Donation proposal under review.
- Slides (01/30/2023 Maintainers meeting).
- Project Roadmap and project Board describing the current state of the project.
- Arrow schemas used by this package.
- Threat model.
This first step is intended to address the specific use cases of traffic reduction. Based on community feedback, many companies want to reduce the cost of transferring telemetry data over the Internet. By adding a collector that acts as a point of integration and traffic conversion at the edge of a client environment, we can take advantage of the columnar format to eliminate redundant data and optimize the compression rate. This is illustrated in the following diagram.
Note 1: A fallback mechanism can be used to handle the case where the new protocol is not supported by the target. More on this mechanism in this section of the OTEP.
The experimental collector implements on top of this library a new OTLP Arrow Receiver and Exporter able to fallback on standard OTLP when needed. The following diagram is an overview of this integration. The internal representation of the data has not been updated and this collector is still fundamentally row-oriented internally.
Note 2: A future phase 2 of this project will focus on implementing end-to-end OTLP arrow to improve the overall performance.
A comparison of the performance and compression rates between the standard OTLP protocol and the OTLP Arrow protocol is given below. Except for the traces, the telemetry data used for these benchmarks are from a synthetic data generator. A larger campaign of tests and benchmarks on real production data will be conducted with the help of the community (see the Help us section).
Current results show an improvement of the compression ratio from 200% to 400%.
The following three tables compare the performance in terms of compression ratio between:
- OTLP,
- OTLP Arrow (streaming mode),
- and OTLP Arrow (unary RPC mode),
for metrics, logs, and traces (in this order).
The following results compare the end-to-end performance between OTLP and OTLP Arrow (streaming mode), and OTLP Arrow (unary RPC mode) for metrics, logs, and traces (in this order).
All the steps are included: encoding, sending, decoding, compression, decompression, and conversion (when that is applicable).
Each result is composed of two tables: one for the phase 1 (current implementation) and one evaluation for the phase 2 (future implementation). The results of phase 2 are extrapolated from the results of phase 1 by removing the conversion steps.
The testing of this package and the validation of the OTLP Arrow encoding/decoding are the object of particular attention because of the central position of this package in the future OTEL collector.
Concerning the tests, the plan is to:
- reach a test coverage close to 80% (currently ~62%).
- add more fuzz tests on the encoding and decoding of OTLP Arrow messages,
- implement integration tests with the experimental collector (testbed implemented in the experimental collector).
Concerning the encoding/decoding validation, the plan is to:
- compare the OTLP entities before and after their conversion to OTLP Arrow entities (done).
- test the conversion procedure of the production data via a CLI tool or directly via the integration in the experimental collector (wip).
A validation of the compression ratio stability is also part of the objectives. This validation will be performed on production data (see Help us section).
We strongly encourage community members to test this protocol extension on production data and report their results and potential problems.
A beta version of this experimental collector will soon be available for this purpose. The testing procedure is described below:
- ToDo
Pull requests are welcome. For major changes, please open an issue first to discuss what you would like to change. For more information, please read CONTRIBUTING.
To (re)generate the ArrowStreamService gRPC service, you need to install the protoc
compiler and the protoc-gen-grpc
plugin.
go install google.golang.org/protobuf/cmd/[email protected]
go install google.golang.org/grpc/cmd/[email protected]
export PATH="$PATH:$(go env GOPATH)/bin"
cd ./proto
./generate.sh
Once the *.pb.go
files are generated, you need to replace the content of the api/collector/arrow/v1
directory by the
generated files present in the ./proto/api/collector/arrow/v1
directory.
The integration of this package with the OpenTelemetry Collector is done in the following experimental repository:
- Project Board
- Milestones:
- Not yet implemented
- Multivariate metrics encoder/decoder
- Complex attributes (list, map)
- Some mitigations mentioned in the thread model assessment
OTEL Arrow Adapter is licensed under Apache 2.0.