-
Notifications
You must be signed in to change notification settings - Fork 2.1k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Thanos Query: gaps in deduplicated data #7656
Comments
Deduplicating time series data is surprisingly hard! I have no great idea how to do it properly. The approach that Thanos takes during query time is roughly that we start with some replica and then if the gap gets too large we switch over. But this had numerous edge cases in the past. I wonder how we could improve it |
Thanks for your reply. I look forward to solving this agonizing problem. |
Yeah I'm happy to brainstorm about this if you have an idea! |
For the moment, I still have no idea where to start and what guidelines we should adopt. except, of course, one consistent data set after deduplication. I wonder if the algorithm for deduplication on Compactor via the “--deduplication.func=penalty” applied to Querier would not solve the problem. Of course, if that's not the cause. |
Penalty is the same algorithm that the querier uses though. |
@MichaHoffmann Wouldn't putting a configurable upper bound on the penalty solve this issue (or at least allow fixing it by configuration)? |
In the sense that we always switch replica if the gap is at least this configured size? |
Sorry, I missed the response.
|
Thanos, Prometheus and Golang version used:
thanos, version 0.35.1 (branch: HEAD, revision: 086a698)
build user: root@be0f036fd8fa
build date: 20240528-13:54:20
go version: go1.21.10
platform: linux/amd64
tags: netgo
prometheus, version 2.32.1 (branch: HEAD, revision: 41f1a8125e664985dd30674e5bdf6b683eff5d32)
build user: root@54b6dbd48b97
build date: 20211217-22:08:06
go version: go1.17.5
platform: linux/amd64
Object Storage Provider:
Ceph
What happened:
Thanos Query: gaps in deduplicated data
What you expected to happen:
Two instances of prometheus scrap data from sources and another federated prometheuses on OpenShift.
As long as we search the data without deduplication, the data is continuous.
Anything else we need to know:
Scereens attached.
What you expected to happen:
Deduplication should properly combine datasets.
The text was updated successfully, but these errors were encountered: