-
Notifications
You must be signed in to change notification settings - Fork 2.1k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Store: smaller block size for snappy decoding #6463
Store: smaller block size for snappy decoding #6463
Conversation
75bd55e
to
1f6643e
Compare
1f6643e
to
9c0476f
Compare
Tests failure seems related. |
9c0476f
to
864c9ad
Compare
fixed and adapted benchmark results! |
864c9ad
to
b3ce522
Compare
pkg/store/postings_codec.go
Outdated
var ( | ||
readersPool = sync.Pool{ | ||
New: func() interface{} { | ||
return s2.NewReader(nil, s2.ReaderMaxBlockSize(4<<10)) |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I thought the other day that perhaps we could not specify a New
function and for gRPC
we could use the default block size whereas for postings we could use the average length of the last 10 inputs, for example. What do you think?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
The problem is that reader max block size needs to be larger than writer block size; so if the average goes down we cannot decode anymore with that approach i think ( i might be wrong )? Using the minimum 4kb feels more traceable to me somehow.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
This change still feels backwards incompatible since i could have cached postings somewhere that were written with larger 64kb block size which might be problematic.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I think the s2 reader ensures that the buffer is enlarged if needed https://github.com/klauspost/compress/blob/master/s2/reader.go#L152 so the only downside AFAICT is that there could be 1 extra allocation. The only complication I see here is that ideally, there should be a way to access the capacity of the internal buffer so that we could pre-allocate a more appropriately sized buffer.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
klauspost/compress#832 trying to push something forward here to allow us better estimate how big the buffer should be
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Ok, in the latest change i added some capacity tracking to the reader pool and use the running average of the capacity of the 10 last returned readers as an estimate for new readers.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Had to revert my changes again, if we dont bound max block sizes the readers will just grow until 64kb again, which would suffer the same issues eventually. I think we can only bound the block sizes.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Or is it because the benchmarks load up a lot of postings whereas the problematic case is about small postings lists?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
That might be the issue with the benchmarks but we dont limit the block size it might easily happen in someones prod too; right?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I think it doesn't matter because the other choice is allocating []byte for the whole decoded response which in this case will be much bigger than https://github.com/google/snappy/blob/main/framing_format.txt#L106. If suddenly smaller postings come in then it's also OK to reuse the readers with bigger reader buffers because they will be freed eventually. The block size is capped in snappy.
b3ce522
to
e3ddddc
Compare
7626751
to
5a077c6
Compare
734d145
to
4a93844
Compare
0fcdfb9
to
d02390a
Compare
d02390a
to
9003bae
Compare
Signed-off-by: Michael Hoffmann <[email protected]>
9003bae
to
251f15d
Compare
Better left to the experts: https://github.com/thanos-io/thanos/pull/6475/files |
wants to address #6434
Changes
extsnappy
to ensure the contract that postings were encoded with a valid block sizeVerification
Benchmarks