Traces have a perpetual problem of being unbelievably useful but extremely expensive because of how many of them can be generated. What if instead of solving this problem by sampling fewer, more interesting traces, you worked on improving the data representation of traces?
In particular, traces today are generic JSON blobs in OpenTelemetry. What if you thought of trace data as a denormalized store where one part is your source code at a particular version, and your traces are short, compact binary formats that point to specific parts of that source code. My intuition is that you could store a much larger number of traces, and perhaps have more efficient queries across them as well. This mainly works if you just want timing information, and don’t care as much about what arguments were passed in. In that case you could store a span with just a few bytes in the following way:
- 2 bytes for the code location - this gives you 65336 places that you can trace in code.
- 4 bytes for timing information - specifically the time it took to run that portion of the trace.
- 4 bytes for a global trace ID, 2 bytes for a trace counter.
With just 12 bytes, you can get a full description of what happened, in what order, and how long it took without storing and processing blobs of JSON all over the place. This could also make tail sampling these traces easier since you can hold more of them in memory.