github.com/cockroachdb/pebble@v0.0.0-20231214172447-ab4952c5f87b/internal/keyspan/fragmenter.go (about) 1 // Copyright 2018 The LevelDB-Go and Pebble Authors. All rights reserved. Use 2 // of this source code is governed by a BSD-style license that can be found in 3 // the LICENSE file. 4 5 package keyspan 6 7 import ( 8 "fmt" 9 "sort" 10 11 "github.com/cockroachdb/pebble/internal/base" 12 "github.com/cockroachdb/pebble/internal/invariants" 13 ) 14 15 type spansByStartKey struct { 16 cmp base.Compare 17 buf []Span 18 } 19 20 func (v *spansByStartKey) Len() int { return len(v.buf) } 21 func (v *spansByStartKey) Less(i, j int) bool { 22 return v.cmp(v.buf[i].Start, v.buf[j].Start) < 0 23 } 24 func (v *spansByStartKey) Swap(i, j int) { 25 v.buf[i], v.buf[j] = v.buf[j], v.buf[i] 26 } 27 28 type spansByEndKey struct { 29 cmp base.Compare 30 buf []Span 31 } 32 33 func (v *spansByEndKey) Len() int { return len(v.buf) } 34 func (v *spansByEndKey) Less(i, j int) bool { 35 return v.cmp(v.buf[i].End, v.buf[j].End) < 0 36 } 37 func (v *spansByEndKey) Swap(i, j int) { 38 v.buf[i], v.buf[j] = v.buf[j], v.buf[i] 39 } 40 41 // keysBySeqNumKind sorts spans by the start key's sequence number in 42 // descending order. If two spans have equal sequence number, they're compared 43 // by key kind in descending order. This ordering matches the ordering of 44 // base.InternalCompare among keys with matching user keys. 45 type keysBySeqNumKind []Key 46 47 func (v *keysBySeqNumKind) Len() int { return len(*v) } 48 func (v *keysBySeqNumKind) Less(i, j int) bool { return (*v)[i].Trailer > (*v)[j].Trailer } 49 func (v *keysBySeqNumKind) Swap(i, j int) { (*v)[i], (*v)[j] = (*v)[j], (*v)[i] } 50 51 // Sort the spans by start key. This is the ordering required by the 52 // Fragmenter. Usually spans are naturally sorted by their start key, 53 // but that isn't true for range deletion tombstones in the legacy 54 // range-del-v1 block format. 55 func Sort(cmp base.Compare, spans []Span) { 56 sorter := spansByStartKey{ 57 cmp: cmp, 58 buf: spans, 59 } 60 sort.Sort(&sorter) 61 } 62 63 // Fragmenter fragments a set of spans such that overlapping spans are 64 // split at their overlap points. The fragmented spans are output to the 65 // supplied Output function. 66 type Fragmenter struct { 67 Cmp base.Compare 68 Format base.FormatKey 69 // Emit is called to emit a fragmented span and its keys. Every key defined 70 // within the emitted Span applies to the entirety of the Span's key span. 71 // Keys are ordered in decreasing order of their sequence numbers, and if 72 // equal, decreasing order of key kind. 73 Emit func(Span) 74 // pending contains the list of pending fragments that have not been 75 // flushed to the block writer. Note that the spans have not been 76 // fragmented on the end keys yet. That happens as the spans are 77 // flushed. All pending spans have the same Start. 78 pending []Span 79 // doneBuf is used to buffer completed span fragments when flushing to a 80 // specific key (e.g. TruncateAndFlushTo). It is cached in the Fragmenter to 81 // allow reuse. 82 doneBuf []Span 83 // sortBuf is used to sort fragments by end key when flushing. 84 sortBuf spansByEndKey 85 // flushBuf is used to sort keys by (seqnum,kind) before emitting. 86 flushBuf keysBySeqNumKind 87 // flushedKey is the key that fragments have been flushed up to. Any 88 // additional spans added to the fragmenter must have a start key >= 89 // flushedKey. A nil value indicates flushedKey has not been set. 90 flushedKey []byte 91 finished bool 92 } 93 94 func (f *Fragmenter) checkInvariants(buf []Span) { 95 for i := 1; i < len(buf); i++ { 96 if f.Cmp(buf[i].Start, buf[i].End) >= 0 { 97 panic(fmt.Sprintf("pebble: empty pending span invariant violated: %s", buf[i])) 98 } 99 if f.Cmp(buf[i-1].Start, buf[i].Start) != 0 { 100 panic(fmt.Sprintf("pebble: pending span invariant violated: %s %s", 101 f.Format(buf[i-1].Start), f.Format(buf[i].Start))) 102 } 103 } 104 } 105 106 // Add adds a span to the fragmenter. Spans may overlap and the 107 // fragmenter will internally split them. The spans must be presented in 108 // increasing start key order. That is, Add must be called with a series 109 // of spans like: 110 // 111 // a---e 112 // c---g 113 // c-----i 114 // j---n 115 // j-l 116 // 117 // We need to fragment the spans at overlap points. In the above 118 // example, we'd create: 119 // 120 // a-c-e 121 // c-e-g 122 // c-e-g-i 123 // j-l-n 124 // j-l 125 // 126 // The fragments need to be output sorted by start key, and for equal start 127 // keys, sorted by descending sequence number. This last part requires a mild 128 // bit of care as the fragments are not created in descending sequence number 129 // order. 130 // 131 // Once a start key has been seen, we know that we'll never see a smaller 132 // start key and can thus flush all of the fragments that lie before that 133 // start key. 134 // 135 // Walking through the example above, we start with: 136 // 137 // a---e 138 // 139 // Next we add [c,g) resulting in: 140 // 141 // a-c-e 142 // c---g 143 // 144 // The fragment [a,c) is flushed leaving the pending spans as: 145 // 146 // c-e 147 // c---g 148 // 149 // The next span is [c,i): 150 // 151 // c-e 152 // c---g 153 // c-----i 154 // 155 // No fragments are flushed. The next span is [j,n): 156 // 157 // c-e 158 // c---g 159 // c-----i 160 // j---n 161 // 162 // The fragments [c,e), [c,g) and [c,i) are flushed. We sort these fragments 163 // by their end key, then split the fragments on the end keys: 164 // 165 // c-e 166 // c-e-g 167 // c-e---i 168 // 169 // The [c,e) fragments all get flushed leaving: 170 // 171 // e-g 172 // e---i 173 // 174 // This process continues until there are no more fragments to flush. 175 // 176 // WARNING: the slices backing Start, End, Keys, Key.Suffix and Key.Value are 177 // all retained after this method returns and should not be modified. This is 178 // safe for spans that are added from a memtable or batch. It is partially 179 // unsafe for a span read from an sstable. Specifically, the Keys slice of a 180 // Span returned during sstable iteration is only valid until the next iterator 181 // operation. The stability of the user keys depend on whether the block is 182 // prefix compressed, and in practice Pebble never prefix compresses range 183 // deletion and range key blocks, so these keys are stable. Because of this key 184 // stability, typically callers only need to perform a shallow clone of the Span 185 // before Add-ing it to the fragmenter. 186 // 187 // Add requires the provided span's keys are sorted in Trailer descending order. 188 func (f *Fragmenter) Add(s Span) { 189 if f.finished { 190 panic("pebble: span fragmenter already finished") 191 } else if s.KeysOrder != ByTrailerDesc { 192 panic("pebble: span keys unexpectedly not in trailer descending order") 193 } 194 if f.flushedKey != nil { 195 switch c := f.Cmp(s.Start, f.flushedKey); { 196 case c < 0: 197 panic(fmt.Sprintf("pebble: start key (%s) < flushed key (%s)", 198 f.Format(s.Start), f.Format(f.flushedKey))) 199 } 200 } 201 if f.Cmp(s.Start, s.End) >= 0 { 202 // An empty span, we can ignore it. 203 return 204 } 205 if invariants.RaceEnabled { 206 f.checkInvariants(f.pending) 207 defer func() { f.checkInvariants(f.pending) }() 208 } 209 210 if len(f.pending) > 0 { 211 // Since all of the pending spans have the same start key, we only need 212 // to compare against the first one. 213 switch c := f.Cmp(f.pending[0].Start, s.Start); { 214 case c > 0: 215 panic(fmt.Sprintf("pebble: keys must be added in order: %s > %s", 216 f.Format(f.pending[0].Start), f.Format(s.Start))) 217 case c == 0: 218 // The new span has the same start key as the existing pending 219 // spans. Add it to the pending buffer. 220 f.pending = append(f.pending, s) 221 return 222 } 223 224 // At this point we know that the new start key is greater than the pending 225 // spans start keys. 226 f.truncateAndFlush(s.Start) 227 } 228 229 f.pending = append(f.pending, s) 230 } 231 232 // Cover is returned by Framenter.Covers and describes a span's relationship to 233 // a key at a particular snapshot. 234 type Cover int8 235 236 const ( 237 // NoCover indicates the tested key does not fall within the span's bounds, 238 // or the span contains no keys with sequence numbers higher than the key's. 239 NoCover Cover = iota 240 // CoversInvisibly indicates the tested key does fall within the span's 241 // bounds and the span contains at least one key with a higher sequence 242 // number, but none visible at the provided snapshot. 243 CoversInvisibly 244 // CoversVisibly indicates the tested key does fall within the span's 245 // bounds, and the span constains at least one key with a sequence number 246 // higher than the key's sequence number that is visible at the provided 247 // snapshot. 248 CoversVisibly 249 ) 250 251 // Covers returns an enum indicating whether the specified key is covered by one 252 // of the pending keys. The provided key must be consistent with the ordering of 253 // the spans. That is, it is invalid to specify a key here that is out of order 254 // with the span start keys passed to Add. 255 func (f *Fragmenter) Covers(key base.InternalKey, snapshot uint64) Cover { 256 if f.finished { 257 panic("pebble: span fragmenter already finished") 258 } 259 if len(f.pending) == 0 { 260 return NoCover 261 } 262 263 if f.Cmp(f.pending[0].Start, key.UserKey) > 0 { 264 panic(fmt.Sprintf("pebble: keys must be in order: %s > %s", 265 f.Format(f.pending[0].Start), key.Pretty(f.Format))) 266 } 267 268 cover := NoCover 269 seqNum := key.SeqNum() 270 for _, s := range f.pending { 271 if f.Cmp(key.UserKey, s.End) < 0 { 272 // NB: A range deletion tombstone does not delete a point operation 273 // at the same sequence number, and broadly a span is not considered 274 // to cover a point operation at the same sequence number. 275 276 for i := range s.Keys { 277 if kseq := s.Keys[i].SeqNum(); kseq > seqNum { 278 // This key from the span has a higher sequence number than 279 // `key`. It covers `key`, although the span's key might not 280 // be visible if its snapshot is too high. 281 // 282 // Batch keys are always be visible. 283 if kseq < snapshot || kseq&base.InternalKeySeqNumBatch != 0 { 284 return CoversVisibly 285 } 286 // s.Keys[i] is not visible. 287 cover = CoversInvisibly 288 } 289 } 290 } 291 } 292 return cover 293 } 294 295 // Empty returns true if all fragments added so far have finished flushing. 296 func (f *Fragmenter) Empty() bool { 297 return f.finished || len(f.pending) == 0 298 } 299 300 // TruncateAndFlushTo flushes all of the fragments with a start key <= key, 301 // truncating spans to the specified end key. Used during compaction to force 302 // emitting of spans which straddle an sstable boundary. Consider 303 // the scenario: 304 // 305 // a---------k#10 306 // f#8 307 // f#7 308 // 309 // Let's say the next user key after f is g. Calling TruncateAndFlushTo(g) will 310 // flush this span: 311 // 312 // a-------g#10 313 // f#8 314 // f#7 315 // 316 // And leave this one in f.pending: 317 // 318 // g----k#10 319 // 320 // WARNING: The fragmenter could hold on to the specified end key. Ensure it's 321 // a safe byte slice that could outlast the current sstable output, and one 322 // that will never be modified. 323 func (f *Fragmenter) TruncateAndFlushTo(key []byte) { 324 if f.finished { 325 panic("pebble: span fragmenter already finished") 326 } 327 if f.flushedKey != nil { 328 switch c := f.Cmp(key, f.flushedKey); { 329 case c < 0: 330 panic(fmt.Sprintf("pebble: start key (%s) < flushed key (%s)", 331 f.Format(key), f.Format(f.flushedKey))) 332 } 333 } 334 if invariants.RaceEnabled { 335 f.checkInvariants(f.pending) 336 defer func() { f.checkInvariants(f.pending) }() 337 } 338 if len(f.pending) > 0 { 339 // Since all of the pending spans have the same start key, we only need 340 // to compare against the first one. 341 switch c := f.Cmp(f.pending[0].Start, key); { 342 case c > 0: 343 panic(fmt.Sprintf("pebble: keys must be added in order: %s > %s", 344 f.Format(f.pending[0].Start), f.Format(key))) 345 case c == 0: 346 return 347 } 348 } 349 f.truncateAndFlush(key) 350 } 351 352 // Start returns the start key of the first span in the pending buffer, or nil 353 // if there are no pending spans. The start key of all pending spans is the same 354 // as that of the first one. 355 func (f *Fragmenter) Start() []byte { 356 if len(f.pending) > 0 { 357 return f.pending[0].Start 358 } 359 return nil 360 } 361 362 // Flushes all pending spans up to key (exclusive). 363 // 364 // WARNING: The specified key is stored without making a copy, so all callers 365 // must ensure it is safe. 366 func (f *Fragmenter) truncateAndFlush(key []byte) { 367 f.flushedKey = append(f.flushedKey[:0], key...) 368 done := f.doneBuf[:0] 369 pending := f.pending 370 f.pending = f.pending[:0] 371 372 // pending and f.pending share the same underlying storage. As we iterate 373 // over pending we append to f.pending, but only one entry is appended in 374 // each iteration, after we have read the entry being overwritten. 375 for _, s := range pending { 376 if f.Cmp(key, s.End) < 0 { 377 // s: a--+--e 378 // new: c------ 379 if f.Cmp(s.Start, key) < 0 { 380 done = append(done, Span{ 381 Start: s.Start, 382 End: key, 383 Keys: s.Keys, 384 }) 385 } 386 f.pending = append(f.pending, Span{ 387 Start: key, 388 End: s.End, 389 Keys: s.Keys, 390 }) 391 } else { 392 // s: a-----e 393 // new: e---- 394 done = append(done, s) 395 } 396 } 397 398 f.doneBuf = done[:0] 399 f.flush(done, nil) 400 } 401 402 // flush a group of range spans to the block. The spans are required to all have 403 // the same start key. We flush all span fragments until startKey > lastKey. If 404 // lastKey is nil, all span fragments are flushed. The specification of a 405 // non-nil lastKey occurs for range deletion tombstones during compaction where 406 // we want to flush (but not truncate) all range tombstones that start at or 407 // before the first key in the next sstable. Consider: 408 // 409 // a---e#10 410 // a------h#9 411 // 412 // If a compaction splits the sstables at key c we want the first sstable to 413 // contain the tombstones [a,e)#10 and [a,e)#9. Fragmentation would naturally 414 // produce a tombstone [e,h)#9, but we don't need to output that tombstone to 415 // the first sstable. 416 func (f *Fragmenter) flush(buf []Span, lastKey []byte) { 417 if invariants.RaceEnabled { 418 f.checkInvariants(buf) 419 } 420 421 // Sort the spans by end key. This will allow us to walk over the spans and 422 // easily determine the next split point (the smallest end-key). 423 f.sortBuf.cmp = f.Cmp 424 f.sortBuf.buf = buf 425 sort.Sort(&f.sortBuf) 426 427 // Loop over the spans, splitting by end key. 428 for len(buf) > 0 { 429 // A prefix of spans will end at split. remove represents the count of 430 // that prefix. 431 remove := 1 432 split := buf[0].End 433 f.flushBuf = append(f.flushBuf[:0], buf[0].Keys...) 434 435 for i := 1; i < len(buf); i++ { 436 if f.Cmp(split, buf[i].End) == 0 { 437 remove++ 438 } 439 f.flushBuf = append(f.flushBuf, buf[i].Keys...) 440 } 441 442 sort.Sort(&f.flushBuf) 443 444 f.Emit(Span{ 445 Start: buf[0].Start, 446 End: split, 447 // Copy the sorted keys to a new slice. 448 // 449 // This allocation is an unfortunate side effect of the Fragmenter and 450 // the expectation that the spans it produces are available in-memory 451 // indefinitely. 452 // 453 // Eventually, we should be able to replace the fragmenter with the 454 // keyspan.MergingIter which will perform just-in-time 455 // fragmentation, and only guaranteeing the memory lifetime for the 456 // current span. The MergingIter fragments while only needing to 457 // access one Span per level. It only accesses the Span at the 458 // current position for each level. During compactions, we can write 459 // these spans to sstables without retaining previous Spans. 460 Keys: append([]Key(nil), f.flushBuf...), 461 }) 462 463 if lastKey != nil && f.Cmp(split, lastKey) > 0 { 464 break 465 } 466 467 // Adjust the start key for every remaining span. 468 buf = buf[remove:] 469 for i := range buf { 470 buf[i].Start = split 471 } 472 } 473 } 474 475 // Finish flushes any remaining fragments to the output. It is an error to call 476 // this if any other spans will be added. 477 func (f *Fragmenter) Finish() { 478 if f.finished { 479 panic("pebble: span fragmenter already finished") 480 } 481 f.flush(f.pending, nil) 482 f.finished = true 483 }