github.com/cockroachdb/pebble@v1.1.2/record/log_writer.go (about) 1 // Copyright 2018 The LevelDB-Go and Pebble Authors. All rights reserved. Use 2 // of this source code is governed by a BSD-style license that can be found in 3 // the LICENSE file. 4 5 package record 6 7 import ( 8 "context" 9 "encoding/binary" 10 "io" 11 "runtime/pprof" 12 "sync" 13 "sync/atomic" 14 "time" 15 16 "github.com/cockroachdb/errors" 17 "github.com/cockroachdb/pebble/internal/base" 18 "github.com/cockroachdb/pebble/internal/crc" 19 "github.com/prometheus/client_golang/prometheus" 20 ) 21 22 var walSyncLabels = pprof.Labels("pebble", "wal-sync") 23 var errClosedWriter = errors.New("pebble/record: closed LogWriter") 24 25 type block struct { 26 // buf[:written] has already been filled with fragments. Updated atomically. 27 written atomic.Int32 28 // buf[:flushed] has already been flushed to w. 29 flushed int32 30 buf [blockSize]byte 31 } 32 33 type flusher interface { 34 Flush() error 35 } 36 37 type syncer interface { 38 Sync() error 39 } 40 41 const ( 42 syncConcurrencyBits = 12 43 44 // SyncConcurrency is the maximum number of concurrent sync operations that 45 // can be performed. Note that a sync operation is initiated either by a call 46 // to SyncRecord or by a call to Close. Exported as this value also limits 47 // the commit concurrency in commitPipeline. 48 SyncConcurrency = 1 << syncConcurrencyBits 49 ) 50 51 type syncSlot struct { 52 wg *sync.WaitGroup 53 err *error 54 } 55 56 // syncQueue is a lock-free fixed-size single-producer, single-consumer 57 // queue. The single-producer can push to the head, and the single-consumer can 58 // pop multiple values from the tail. Popping calls Done() on each of the 59 // available *sync.WaitGroup elements. 60 type syncQueue struct { 61 // headTail packs together a 32-bit head index and a 32-bit tail index. Both 62 // are indexes into slots modulo len(slots)-1. 63 // 64 // tail = index of oldest data in queue 65 // head = index of next slot to fill 66 // 67 // Slots in the range [tail, head) are owned by consumers. A consumer 68 // continues to own a slot outside this range until it nils the slot, at 69 // which point ownership passes to the producer. 70 // 71 // The head index is stored in the most-significant bits so that we can 72 // atomically add to it and the overflow is harmless. 73 headTail atomic.Uint64 74 75 // slots is a ring buffer of values stored in this queue. The size must be a 76 // power of 2. A slot is in use until the tail index has moved beyond it. 77 slots [SyncConcurrency]syncSlot 78 79 // blocked is an atomic boolean which indicates whether syncing is currently 80 // blocked or can proceed. It is used by the implementation of 81 // min-sync-interval to block syncing until the min interval has passed. 82 blocked atomic.Bool 83 } 84 85 const dequeueBits = 32 86 87 func (q *syncQueue) unpack(ptrs uint64) (head, tail uint32) { 88 const mask = 1<<dequeueBits - 1 89 head = uint32((ptrs >> dequeueBits) & mask) 90 tail = uint32(ptrs & mask) 91 return 92 } 93 94 func (q *syncQueue) push(wg *sync.WaitGroup, err *error) { 95 ptrs := q.headTail.Load() 96 head, tail := q.unpack(ptrs) 97 if (tail+uint32(len(q.slots)))&(1<<dequeueBits-1) == head { 98 panic("pebble: queue is full") 99 } 100 101 slot := &q.slots[head&uint32(len(q.slots)-1)] 102 slot.wg = wg 103 slot.err = err 104 105 // Increment head. This passes ownership of slot to dequeue and acts as a 106 // store barrier for writing the slot. 107 q.headTail.Add(1 << dequeueBits) 108 } 109 110 func (q *syncQueue) setBlocked() { 111 q.blocked.Store(true) 112 } 113 114 func (q *syncQueue) clearBlocked() { 115 q.blocked.Store(false) 116 } 117 118 func (q *syncQueue) empty() bool { 119 head, tail, _ := q.load() 120 return head == tail 121 } 122 123 // load returns the head, tail of the queue for what should be synced to the 124 // caller. It can return a head, tail of zero if syncing is blocked due to 125 // min-sync-interval. It additionally returns the real length of this queue, 126 // regardless of whether syncing is blocked. 127 func (q *syncQueue) load() (head, tail, realLength uint32) { 128 ptrs := q.headTail.Load() 129 head, tail = q.unpack(ptrs) 130 realLength = head - tail 131 if q.blocked.Load() { 132 return 0, 0, realLength 133 } 134 return head, tail, realLength 135 } 136 137 // REQUIRES: queueSemChan is non-nil. 138 func (q *syncQueue) pop(head, tail uint32, err error, queueSemChan chan struct{}) error { 139 if tail == head { 140 // Queue is empty. 141 return nil 142 } 143 144 for ; tail != head; tail++ { 145 slot := &q.slots[tail&uint32(len(q.slots)-1)] 146 wg := slot.wg 147 if wg == nil { 148 return errors.Errorf("nil waiter at %d", errors.Safe(tail&uint32(len(q.slots)-1))) 149 } 150 *slot.err = err 151 slot.wg = nil 152 slot.err = nil 153 // We need to bump the tail count before signalling the wait group as 154 // signalling the wait group can trigger release a blocked goroutine which 155 // will try to enqueue before we've "freed" space in the queue. 156 q.headTail.Add(1) 157 wg.Done() 158 // Is always non-nil in production. 159 if queueSemChan != nil { 160 <-queueSemChan 161 } 162 } 163 164 return nil 165 } 166 167 // flusherCond is a specialized condition variable that allows its condition to 168 // change and readiness be signalled without holding its associated mutex. In 169 // particular, when a waiter is added to syncQueue atomically, this condition 170 // variable can be signalled without holding flusher.Mutex. 171 type flusherCond struct { 172 mu *sync.Mutex 173 q *syncQueue 174 cond sync.Cond 175 } 176 177 func (c *flusherCond) init(mu *sync.Mutex, q *syncQueue) { 178 c.mu = mu 179 c.q = q 180 // Yes, this is a bit circular, but that is intentional. flusherCond.cond.L 181 // points flusherCond so that when cond.L.Unlock is called flusherCond.Unlock 182 // will be called and we can check the !syncQueue.empty() condition. 183 c.cond.L = c 184 } 185 186 func (c *flusherCond) Signal() { 187 // Pass-through to the cond var. 188 c.cond.Signal() 189 } 190 191 func (c *flusherCond) Wait() { 192 // Pass-through to the cond var. Note that internally the cond var implements 193 // Wait as: 194 // 195 // t := notifyListAdd() 196 // L.Unlock() 197 // notifyListWait(t) 198 // L.Lock() 199 // 200 // We've configured the cond var to call flusherReady.Unlock() which allows 201 // us to check the !syncQueue.empty() condition without a danger of missing a 202 // notification. Any call to flusherReady.Signal() after notifyListAdd() is 203 // called will cause the subsequent notifyListWait() to return immediately. 204 c.cond.Wait() 205 } 206 207 func (c *flusherCond) Lock() { 208 c.mu.Lock() 209 } 210 211 func (c *flusherCond) Unlock() { 212 c.mu.Unlock() 213 if !c.q.empty() { 214 // If the current goroutine is about to block on sync.Cond.Wait, this call 215 // to Signal will prevent that. The comment in Wait above explains a bit 216 // about what is going on here, but it is worth reiterating: 217 // 218 // flusherCond.Wait() 219 // sync.Cond.Wait() 220 // t := notifyListAdd() 221 // flusherCond.Unlock() <-- we are here 222 // notifyListWait(t) 223 // flusherCond.Lock() 224 // 225 // The call to Signal here results in: 226 // 227 // sync.Cond.Signal() 228 // notifyListNotifyOne() 229 // 230 // The call to notifyListNotifyOne() will prevent the call to 231 // notifyListWait(t) from blocking. 232 c.cond.Signal() 233 } 234 } 235 236 type durationFunc func() time.Duration 237 238 // syncTimer is an interface for timers, modeled on the closure callback mode 239 // of time.Timer. See time.AfterFunc and LogWriter.afterFunc. syncTimer is used 240 // by tests to mock out the timer functionality used to implement 241 // min-sync-interval. 242 type syncTimer interface { 243 Reset(time.Duration) bool 244 Stop() bool 245 } 246 247 // LogWriter writes records to an underlying io.Writer. In order to support WAL 248 // file reuse, a LogWriter's records are tagged with the WAL's file 249 // number. When reading a log file a record from a previous incarnation of the 250 // file will return the error ErrInvalidLogNum. 251 type LogWriter struct { 252 // w is the underlying writer. 253 w io.Writer 254 // c is w as a closer. 255 c io.Closer 256 // s is w as a syncer. 257 s syncer 258 // logNum is the low 32-bits of the log's file number. 259 logNum uint32 260 // blockNum is the zero based block number for the current block. 261 blockNum int64 262 // err is any accumulated error. TODO(peter): This needs to be protected in 263 // some fashion. Perhaps using atomic.Value. 264 err error 265 // block is the current block being written. Protected by flusher.Mutex. 266 block *block 267 free struct { 268 sync.Mutex 269 blocks []*block 270 } 271 272 flusher struct { 273 sync.Mutex 274 // Flusher ready is a condition variable that is signalled when there are 275 // blocks to flush, syncing has been requested, or the LogWriter has been 276 // closed. For signalling of a sync, it is safe to call without holding 277 // flusher.Mutex. 278 ready flusherCond 279 // Set to true when the flush loop should be closed. 280 close bool 281 // Closed when the flush loop has terminated. 282 closed chan struct{} 283 // Accumulated flush error. 284 err error 285 // minSyncInterval is the minimum duration between syncs. 286 minSyncInterval durationFunc 287 fsyncLatency prometheus.Histogram 288 pending []*block 289 syncQ syncQueue 290 metrics *LogWriterMetrics 291 } 292 293 // afterFunc is a hook to allow tests to mock out the timer functionality 294 // used for min-sync-interval. In normal operation this points to 295 // time.AfterFunc. 296 afterFunc func(d time.Duration, f func()) syncTimer 297 298 // See the comment for LogWriterConfig.QueueSemChan. 299 queueSemChan chan struct{} 300 } 301 302 // LogWriterConfig is a struct used for configuring new LogWriters 303 type LogWriterConfig struct { 304 WALMinSyncInterval durationFunc 305 WALFsyncLatency prometheus.Histogram 306 // QueueSemChan is an optional channel to pop from when popping from 307 // LogWriter.flusher.syncQueue. It functions as a semaphore that prevents 308 // the syncQueue from overflowing (which will cause a panic). All production 309 // code ensures this is non-nil. 310 QueueSemChan chan struct{} 311 } 312 313 // initialAllocatedBlocksCap is the initial capacity of the various slices 314 // intended to hold LogWriter blocks. The LogWriter may allocate more blocks 315 // than this threshold allows. 316 const initialAllocatedBlocksCap = 32 317 318 // blockPool pools *blocks to avoid allocations. Blocks are only added to the 319 // Pool when a LogWriter is closed. Before that, free blocks are maintained 320 // within a LogWriter's own internal free list `w.free.blocks`. 321 var blockPool = sync.Pool{ 322 New: func() any { return &block{} }, 323 } 324 325 // NewLogWriter returns a new LogWriter. 326 func NewLogWriter(w io.Writer, logNum base.FileNum, logWriterConfig LogWriterConfig) *LogWriter { 327 c, _ := w.(io.Closer) 328 s, _ := w.(syncer) 329 r := &LogWriter{ 330 w: w, 331 c: c, 332 s: s, 333 // NB: we truncate the 64-bit log number to 32-bits. This is ok because a) 334 // we are very unlikely to reach a file number of 4 billion and b) the log 335 // number is used as a validation check and using only the low 32-bits is 336 // sufficient for that purpose. 337 logNum: uint32(logNum), 338 afterFunc: func(d time.Duration, f func()) syncTimer { 339 return time.AfterFunc(d, f) 340 }, 341 queueSemChan: logWriterConfig.QueueSemChan, 342 } 343 r.free.blocks = make([]*block, 0, initialAllocatedBlocksCap) 344 r.block = blockPool.Get().(*block) 345 r.flusher.ready.init(&r.flusher.Mutex, &r.flusher.syncQ) 346 r.flusher.closed = make(chan struct{}) 347 r.flusher.pending = make([]*block, 0, cap(r.free.blocks)) 348 r.flusher.metrics = &LogWriterMetrics{} 349 350 f := &r.flusher 351 f.minSyncInterval = logWriterConfig.WALMinSyncInterval 352 f.fsyncLatency = logWriterConfig.WALFsyncLatency 353 354 go func() { 355 pprof.Do(context.Background(), walSyncLabels, r.flushLoop) 356 }() 357 return r 358 } 359 360 func (w *LogWriter) flushLoop(context.Context) { 361 f := &w.flusher 362 f.Lock() 363 364 // Initialize idleStartTime to when the loop starts. 365 idleStartTime := time.Now() 366 var syncTimer syncTimer 367 defer func() { 368 // Capture the idle duration between the last piece of work and when the 369 // loop terminated. 370 f.metrics.WriteThroughput.IdleDuration += time.Since(idleStartTime) 371 if syncTimer != nil { 372 syncTimer.Stop() 373 } 374 close(f.closed) 375 f.Unlock() 376 }() 377 378 // The flush loop performs flushing of full and partial data blocks to the 379 // underlying writer (LogWriter.w), syncing of the writer, and notification 380 // to sync requests that they have completed. 381 // 382 // - flusher.ready is a condition variable that is signalled when there is 383 // work to do. Full blocks are contained in flusher.pending. The current 384 // partial block is in LogWriter.block. And sync operations are held in 385 // flusher.syncQ. 386 // 387 // - The decision to sync is determined by whether there are any sync 388 // requests present in flusher.syncQ and whether enough time has elapsed 389 // since the last sync. If not enough time has elapsed since the last sync, 390 // flusher.syncQ.blocked will be set to 1. If syncing is blocked, 391 // syncQueue.empty() will return true and syncQueue.load() will return 0,0 392 // (i.e. an empty list). 393 // 394 // - flusher.syncQ.blocked is cleared by a timer that is initialized when 395 // blocked is set to 1. When blocked is 1, no syncing will take place, but 396 // flushing will continue to be performed. The on/off toggle for syncing 397 // does not need to be carefully synchronized with the rest of processing 398 // -- all we need to ensure is that after any transition to blocked=1 there 399 // is eventually a transition to blocked=0. syncTimer performs this 400 // transition. Note that any change to min-sync-interval will not take 401 // effect until the previous timer elapses. 402 // 403 // - Picking up the syncing work to perform requires coordination with 404 // picking up the flushing work. Specifically, flushing work is queued 405 // before syncing work. The guarantee of this code is that when a sync is 406 // requested, any previously queued flush work will be synced. This 407 // motivates reading the syncing work (f.syncQ.load()) before picking up 408 // the flush work (w.block.written.Load()). 409 410 // The list of full blocks that need to be written. This is copied from 411 // f.pending on every loop iteration, though the number of elements is 412 // usually small (most frequently 1). In the case of the WAL LogWriter, the 413 // number of blocks is bounded by the size of the WAL's corresponding 414 // memtable (MemtableSize/BlockSize). With the default 64 MiB memtables, 415 // this works out to at most 2048 elements if the entirety of the memtable's 416 // contents are queued. 417 pending := make([]*block, 0, cap(f.pending)) 418 for { 419 for { 420 // Grab the portion of the current block that requires flushing. Note that 421 // the current block can be added to the pending blocks list after we release 422 // the flusher lock, but it won't be part of pending. 423 written := w.block.written.Load() 424 if len(f.pending) > 0 || written > w.block.flushed || !f.syncQ.empty() { 425 break 426 } 427 if f.close { 428 // If the writer is closed, pretend the sync timer fired immediately so 429 // that we can process any queued sync requests. 430 f.syncQ.clearBlocked() 431 if !f.syncQ.empty() { 432 break 433 } 434 return 435 } 436 f.ready.Wait() 437 continue 438 } 439 // Found work to do, so no longer idle. 440 workStartTime := time.Now() 441 idleDuration := workStartTime.Sub(idleStartTime) 442 pending = append(pending[:0], f.pending...) 443 f.pending = f.pending[:0] 444 f.metrics.PendingBufferLen.AddSample(int64(len(pending))) 445 446 // Grab the list of sync waiters. Note that syncQueue.load() will return 447 // 0,0 while we're waiting for the min-sync-interval to expire. This 448 // allows flushing to proceed even if we're not ready to sync. 449 head, tail, realSyncQLen := f.syncQ.load() 450 f.metrics.SyncQueueLen.AddSample(int64(realSyncQLen)) 451 452 // Grab the portion of the current block that requires flushing. Note that 453 // the current block can be added to the pending blocks list after we 454 // release the flusher lock, but it won't be part of pending. This has to 455 // be ordered after we get the list of sync waiters from syncQ in order to 456 // prevent a race where a waiter adds itself to syncQ, but this thread 457 // picks up the entry in syncQ and not the buffered data. 458 written := w.block.written.Load() 459 data := w.block.buf[w.block.flushed:written] 460 w.block.flushed = written 461 462 // If flusher has an error, we propagate it to waiters. Note in spite of 463 // error we consume the pending list above to free blocks for writers. 464 if f.err != nil { 465 f.syncQ.pop(head, tail, f.err, w.queueSemChan) 466 // Update the idleStartTime if work could not be done, so that we don't 467 // include the duration we tried to do work as idle. We don't bother 468 // with the rest of the accounting, which means we will undercount. 469 idleStartTime = time.Now() 470 continue 471 } 472 f.Unlock() 473 synced, syncLatency, bytesWritten, err := w.flushPending(data, pending, head, tail) 474 f.Lock() 475 if synced && f.fsyncLatency != nil { 476 f.fsyncLatency.Observe(float64(syncLatency)) 477 } 478 f.err = err 479 if f.err != nil { 480 f.syncQ.clearBlocked() 481 // Update the idleStartTime if work could not be done, so that we don't 482 // include the duration we tried to do work as idle. We don't bother 483 // with the rest of the accounting, which means we will undercount. 484 idleStartTime = time.Now() 485 continue 486 } 487 488 if synced && f.minSyncInterval != nil { 489 // A sync was performed. Make sure we've waited for the min sync 490 // interval before syncing again. 491 if min := f.minSyncInterval(); min > 0 { 492 f.syncQ.setBlocked() 493 if syncTimer == nil { 494 syncTimer = w.afterFunc(min, func() { 495 f.syncQ.clearBlocked() 496 f.ready.Signal() 497 }) 498 } else { 499 syncTimer.Reset(min) 500 } 501 } 502 } 503 // Finished work, and started idling. 504 idleStartTime = time.Now() 505 workDuration := idleStartTime.Sub(workStartTime) 506 f.metrics.WriteThroughput.Bytes += bytesWritten 507 f.metrics.WriteThroughput.WorkDuration += workDuration 508 f.metrics.WriteThroughput.IdleDuration += idleDuration 509 } 510 } 511 512 func (w *LogWriter) flushPending( 513 data []byte, pending []*block, head, tail uint32, 514 ) (synced bool, syncLatency time.Duration, bytesWritten int64, err error) { 515 defer func() { 516 // Translate panics into errors. The errors will cause flushLoop to shut 517 // down, but allows us to do so in a controlled way and avoid swallowing 518 // the stack that created the panic if panic'ing itself hits a panic 519 // (e.g. unlock of unlocked mutex). 520 if r := recover(); r != nil { 521 err = errors.Newf("%v", r) 522 } 523 }() 524 525 for _, b := range pending { 526 bytesWritten += blockSize - int64(b.flushed) 527 if err = w.flushBlock(b); err != nil { 528 break 529 } 530 } 531 if n := len(data); err == nil && n > 0 { 532 bytesWritten += int64(n) 533 _, err = w.w.Write(data) 534 } 535 536 synced = head != tail 537 if synced { 538 if err == nil && w.s != nil { 539 syncLatency, err = w.syncWithLatency() 540 } 541 f := &w.flusher 542 if popErr := f.syncQ.pop(head, tail, err, w.queueSemChan); popErr != nil { 543 return synced, syncLatency, bytesWritten, popErr 544 } 545 } 546 547 return synced, syncLatency, bytesWritten, err 548 } 549 550 func (w *LogWriter) syncWithLatency() (time.Duration, error) { 551 start := time.Now() 552 err := w.s.Sync() 553 syncLatency := time.Since(start) 554 return syncLatency, err 555 } 556 557 func (w *LogWriter) flushBlock(b *block) error { 558 if _, err := w.w.Write(b.buf[b.flushed:]); err != nil { 559 return err 560 } 561 b.written.Store(0) 562 b.flushed = 0 563 w.free.Lock() 564 w.free.blocks = append(w.free.blocks, b) 565 w.free.Unlock() 566 return nil 567 } 568 569 // queueBlock queues the current block for writing to the underlying writer, 570 // allocates a new block and reserves space for the next header. 571 func (w *LogWriter) queueBlock() { 572 // Allocate a new block, blocking until one is available. We do this first 573 // because w.block is protected by w.flusher.Mutex. 574 w.free.Lock() 575 if len(w.free.blocks) == 0 { 576 w.free.blocks = append(w.free.blocks, blockPool.Get().(*block)) 577 } 578 nextBlock := w.free.blocks[len(w.free.blocks)-1] 579 w.free.blocks = w.free.blocks[:len(w.free.blocks)-1] 580 w.free.Unlock() 581 582 f := &w.flusher 583 f.Lock() 584 f.pending = append(f.pending, w.block) 585 w.block = nextBlock 586 f.ready.Signal() 587 w.err = w.flusher.err 588 f.Unlock() 589 590 w.blockNum++ 591 } 592 593 // Close flushes and syncs any unwritten data and closes the writer. 594 // Where required, external synchronisation is provided by commitPipeline.mu. 595 func (w *LogWriter) Close() error { 596 f := &w.flusher 597 598 // Emit an EOF trailer signifying the end of this log. This helps readers 599 // differentiate between a corrupted entry in the middle of a log from 600 // garbage at the tail from a recycled log file. 601 w.emitEOFTrailer() 602 603 // Signal the flush loop to close. 604 f.Lock() 605 f.close = true 606 f.ready.Signal() 607 f.Unlock() 608 609 // Wait for the flush loop to close. The flush loop will not close until all 610 // pending data has been written or an error occurs. 611 <-f.closed 612 613 // Sync any flushed data to disk. NB: flushLoop will sync after flushing the 614 // last buffered data only if it was requested via syncQ, so we need to sync 615 // here to ensure that all the data is synced. 616 err := w.flusher.err 617 var syncLatency time.Duration 618 if err == nil && w.s != nil { 619 syncLatency, err = w.syncWithLatency() 620 } 621 f.Lock() 622 if f.fsyncLatency != nil { 623 f.fsyncLatency.Observe(float64(syncLatency)) 624 } 625 free := w.free.blocks 626 f.Unlock() 627 628 if w.c != nil { 629 cerr := w.c.Close() 630 w.c = nil 631 if cerr != nil { 632 return cerr 633 } 634 } 635 636 for _, b := range free { 637 b.flushed = 0 638 b.written.Store(0) 639 blockPool.Put(b) 640 } 641 642 w.err = errClosedWriter 643 return err 644 } 645 646 // WriteRecord writes a complete record. Returns the offset just past the end 647 // of the record. 648 // External synchronisation provided by commitPipeline.mu. 649 func (w *LogWriter) WriteRecord(p []byte) (int64, error) { 650 logSize, err := w.SyncRecord(p, nil, nil) 651 return logSize, err 652 } 653 654 // SyncRecord writes a complete record. If wg != nil the record will be 655 // asynchronously persisted to the underlying writer and done will be called on 656 // the wait group upon completion. Returns the offset just past the end of the 657 // record. 658 // External synchronisation provided by commitPipeline.mu. 659 func (w *LogWriter) SyncRecord( 660 p []byte, wg *sync.WaitGroup, err *error, 661 ) (logSize int64, err2 error) { 662 if w.err != nil { 663 return -1, w.err 664 } 665 666 // The `i == 0` condition ensures we handle empty records. Such records can 667 // possibly be generated for VersionEdits stored in the MANIFEST. While the 668 // MANIFEST is currently written using Writer, it is good to support the same 669 // semantics with LogWriter. 670 for i := 0; i == 0 || len(p) > 0; i++ { 671 p = w.emitFragment(i, p) 672 } 673 674 if wg != nil { 675 // If we've been asked to persist the record, add the WaitGroup to the sync 676 // queue and signal the flushLoop. Note that flushLoop will write partial 677 // blocks to the file if syncing has been requested. The contract is that 678 // any record written to the LogWriter to this point will be flushed to the 679 // OS and synced to disk. 680 f := &w.flusher 681 f.syncQ.push(wg, err) 682 f.ready.Signal() 683 } 684 685 offset := w.blockNum*blockSize + int64(w.block.written.Load()) 686 // Note that we don't return w.err here as a concurrent call to Close would 687 // race with our read. That's ok because the only error we could be seeing is 688 // one to syncing for which the caller can receive notification of by passing 689 // in a non-nil err argument. 690 return offset, nil 691 } 692 693 // Size returns the current size of the file. 694 // External synchronisation provided by commitPipeline.mu. 695 func (w *LogWriter) Size() int64 { 696 return w.blockNum*blockSize + int64(w.block.written.Load()) 697 } 698 699 func (w *LogWriter) emitEOFTrailer() { 700 // Write a recyclable chunk header with a different log number. Readers 701 // will treat the header as EOF when the log number does not match. 702 b := w.block 703 i := b.written.Load() 704 binary.LittleEndian.PutUint32(b.buf[i+0:i+4], 0) // CRC 705 binary.LittleEndian.PutUint16(b.buf[i+4:i+6], 0) // Size 706 b.buf[i+6] = recyclableFullChunkType 707 binary.LittleEndian.PutUint32(b.buf[i+7:i+11], w.logNum+1) // Log number 708 b.written.Store(i + int32(recyclableHeaderSize)) 709 } 710 711 func (w *LogWriter) emitFragment(n int, p []byte) (remainingP []byte) { 712 b := w.block 713 i := b.written.Load() 714 first := n == 0 715 last := blockSize-i-recyclableHeaderSize >= int32(len(p)) 716 717 if last { 718 if first { 719 b.buf[i+6] = recyclableFullChunkType 720 } else { 721 b.buf[i+6] = recyclableLastChunkType 722 } 723 } else { 724 if first { 725 b.buf[i+6] = recyclableFirstChunkType 726 } else { 727 b.buf[i+6] = recyclableMiddleChunkType 728 } 729 } 730 731 binary.LittleEndian.PutUint32(b.buf[i+7:i+11], w.logNum) 732 733 r := copy(b.buf[i+recyclableHeaderSize:], p) 734 j := i + int32(recyclableHeaderSize+r) 735 binary.LittleEndian.PutUint32(b.buf[i+0:i+4], crc.New(b.buf[i+6:j]).Value()) 736 binary.LittleEndian.PutUint16(b.buf[i+4:i+6], uint16(r)) 737 b.written.Store(j) 738 739 if blockSize-b.written.Load() < recyclableHeaderSize { 740 // There is no room for another fragment in the block, so fill the 741 // remaining bytes with zeros and queue the block for flushing. 742 for i := b.written.Load(); i < blockSize; i++ { 743 b.buf[i] = 0 744 } 745 w.queueBlock() 746 } 747 return p[r:] 748 } 749 750 // Metrics must be called after Close. The callee will no longer modify the 751 // returned LogWriterMetrics. 752 func (w *LogWriter) Metrics() *LogWriterMetrics { 753 return w.flusher.metrics 754 } 755 756 // LogWriterMetrics contains misc metrics for the log writer. 757 type LogWriterMetrics struct { 758 WriteThroughput base.ThroughputMetric 759 PendingBufferLen base.GaugeSampleMetric 760 SyncQueueLen base.GaugeSampleMetric 761 } 762 763 // Merge merges metrics from x. Requires that x is non-nil. 764 func (m *LogWriterMetrics) Merge(x *LogWriterMetrics) error { 765 m.WriteThroughput.Merge(x.WriteThroughput) 766 m.PendingBufferLen.Merge(x.PendingBufferLen) 767 m.SyncQueueLen.Merge(x.SyncQueueLen) 768 return nil 769 }