Skip to content

feat(profiles): rollout data compression #92133

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Merged
merged 1 commit into from
May 22, 2025
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
7 changes: 7 additions & 0 deletions src/sentry/options/defaults.py
Original file line number Diff line number Diff line change
Expand Up @@ -3113,6 +3113,13 @@
flags=FLAG_AUTOMATOR_MODIFIABLE,
)

register(
"taskworker.try_compress.profile_metrics.rollout",
default=0.0,
type=Float,
flags=FLAG_AUTOMATOR_MODIFIABLE,
)

# Taskbroker flags
register(
"taskworker.try_compress.profile_metrics.level",
Expand Down
31 changes: 19 additions & 12 deletions src/sentry/profiles/consumers/process/factory.py
Original file line number Diff line number Diff line change
Expand Up @@ -18,32 +18,39 @@ def process_message(message: Message[KafkaPayload]) -> None:
sampled = is_sampled(message.payload.headers)

if sampled or options.get("profiling.profile_metrics.unsampled_profiles.enabled"):
b64encoded = b64encode(message.payload.value).decode("utf-8")
process_profile_task.delay(payload=b64encoded, sampled=sampled, compressed_profile=False)
b64encoded_uncompressed = b64encode(message.payload.value).decode("utf-8")

if random.random() < options.get("taskworker.try_compress.profile_metrics"):
if random.random() < options.get("taskworker.try_compress.profile_metrics.rollout"):
Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

No need to change this, but in the future you can use sentry.options.rollout.in_random_rollout() for these kinds of checks.

Copy link
Member Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Ah that's good to know, thank you!

import time
import zlib

metrics.distribution("profiling.profile_metrics.uncompressed_bytes", len(b64encoded))
metrics.distribution(
"profiling.profile_metrics.uncompressed_bytes", len(b64encoded_uncompressed)
)

start_time = time.perf_counter()
b64encoded_compressed = b64encode(
zlib.compress(
message.payload.value,
Comment on lines +32 to +34
Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

sorry for the drive-by review after the fact:

Is there a specific reason why you went with zlib over zstd, which is an overall better compression algorithm which should be preferred?
And is the base64 encoding a hard requirement because the tasks can’t handle bytes arguments?

It is just really weird that this end up being base64(zlib(base64(msgpack))).
We could just make this zstd(msgpack), or base64(zstd(msgpack)) in case we really can’t have bytes.

base64 inflates the payload size by 33% by definition, and its a bit wasteful to do it twice even.

Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Is there a specific reason why you went with zlib over zstd, which is an overall better compression algorithm which should be preferred?

zlib was in stdlib, and I didn't know we already had the necessary dependencies for zstandard.

It is just really weird that this end up being base64(zlib(base64(msgpack))).

The current implementation doesn't have a double base64 encode. We do base64 twice so that we can measure the effects of compression, but the task payload should only be base64 encoded once after compression.

And is the base64 encoding a hard requirement because the tasks can’t handle bytes arguments?

Yes, bytes aren't JSON encodable, so we needed a way to get a str.

Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

thanks for clarifying. I got confused by the double-base64, re-reading this again, its clear there is no double-encoding going on :-D

level=options.get("taskworker.try_compress.profile_metrics.level"),
)
)
metrics.distribution(
"profiling.profile_metrics.compressed_bytes",
len(
b64encode(
zlib.compress(
message.payload.value,
level=options.get("taskworker.try_compress.profile_metrics.level"),
)
)
),
len(b64encoded_compressed),
)
end_time = time.perf_counter()
metrics.distribution(
"profiling.profile_metrics.compression_time",
end_time - start_time,
)
process_profile_task.delay(
payload=b64encoded_compressed, sampled=sampled, compressed_profile=True
)
else:
process_profile_task.delay(
payload=b64encoded_uncompressed, sampled=sampled, compressed_profile=False
)


class ProcessProfileStrategyFactory(ProcessingStrategyFactory[KafkaPayload]):
Expand Down
8 changes: 4 additions & 4 deletions tests/sentry/processing/backpressure/test_checking.py
Original file line number Diff line number Diff line change
Expand Up @@ -30,7 +30,7 @@
"backpressure.checking.interval": 5,
"backpressure.monitoring.enabled": True,
"backpressure.status_ttl": 60,
"taskworker.try_compress.profile_metrics": False,
"taskworker.try_compress.profile_metrics.rollout": 0,
}
)
def test_backpressure_unhealthy_profiles():
Expand All @@ -54,7 +54,7 @@ def test_backpressure_unhealthy_profiles():
"backpressure.checking.interval": 5,
"backpressure.monitoring.enabled": False,
"backpressure.status_ttl": 60,
"taskworker.try_compress.profile_metrics": False,
"taskworker.try_compress.profile_metrics.rollout": 0,
}
)
def test_bad_config():
Expand All @@ -69,7 +69,7 @@ def test_bad_config():
"backpressure.checking.interval": 5,
"backpressure.monitoring.enabled": True,
"backpressure.status_ttl": 60,
"taskworker.try_compress.profile_metrics": False,
"taskworker.try_compress.profile_metrics.rollout": 0,
}
)
def test_backpressure_healthy_profiles(process_profile_task):
Expand Down Expand Up @@ -141,7 +141,7 @@ def test_backpressure_healthy_events(preprocess_event):
{
"backpressure.checking.enabled": False,
"backpressure.checking.interval": 5,
"taskworker.try_compress.profile_metrics": False,
"taskworker.try_compress.profile_metrics.rollout": 0,
}
)
def test_backpressure_not_enabled(process_profile_task):
Expand Down
Loading