github.com/keybase/client/go@v0.0.0-20241007131713-f10651d043c8/kbfs/data/dirty_bcache.go (about) 1 // Copyright 2016 Keybase Inc. All rights reserved. 2 // Use of this source code is governed by a BSD 3 // license that can be found in the LICENSE file. 4 5 package data 6 7 import ( 8 "fmt" 9 "sync" 10 "time" 11 12 "github.com/keybase/client/go/kbfs/idutil" 13 "github.com/keybase/client/go/kbfs/kbfsblock" 14 "github.com/keybase/client/go/kbfs/tlf" 15 "github.com/keybase/client/go/libkb" 16 "github.com/keybase/client/go/logger" 17 "golang.org/x/net/context" 18 ) 19 20 type dirtyBlockID struct { 21 id kbfsblock.ID 22 refNonce kbfsblock.RefNonce 23 branch BranchName 24 } 25 26 type dirtyReq struct { 27 respChan chan<- struct{} 28 bytes int64 29 start time.Time 30 deadline time.Time 31 } 32 33 const ( 34 resetBufferCapTimeDefault = 5 * time.Minute 35 ) 36 37 // DirtyBlockCacheStandard implements the DirtyBlockCache interface by 38 // storing blocks in an in-memory cache. Dirty blocks are identified 39 // by their block ID, branch name, and reference nonce, since the same 40 // block may be forked and modified on different branches and under 41 // different references simultaneously. 42 // 43 // DirtyBlockCacheStandard controls how fast uses can write into KBFS, 44 // and does so with a TCP-like slow-start algorithm that adjusts 45 // itself according to how fast bytes are synced to the server. 46 // Conceptually, there are two buffers: 47 // 48 // syncBuf: The bytes that are currently syncing, or have finished 49 // syncing, back to the servers. Each TLF has only one sync at a 50 // time, but multiple TLFs may be syncing at the same time. We also 51 // track how many bytes within this buffer have finished syncing. 52 // 53 // waitBuf: The bytes that have not yet begun syncing to the 54 // servers. Again, this can be for multiple TLFs, and from multiple 55 // files within a TLF. In the TCP analogy, think of this as the 56 // congestion window (cwnd). 57 // 58 // The goal is to make sure that syncBuf can always be transmitted to 59 // the server within the file system operation timeout forced on us by 60 // the layer that interacts with the file system (19 seconds on OS X 61 // and Windows, defaults to 30 seconds for other layers if not already 62 // set). In fact, ideally the data would be transmitted in HALF of 63 // the file system operation timeout, in case a user Sync operation 64 // gets blocked behind a background Sync operation when there is 65 // significant data in waitBuf. At the same time, we want it to be as 66 // big as possible, because we get the best performance when writing 67 // lots of blocks in parallel to the servers. So, we want an 68 // algorithm that allows waitBuf to grow, without causing the next 69 // sync (or write, or setattr, etc) operation to timeout. For the 70 // purposes of this discussion, let's assume there is only one active 71 // TLF at a time. 72 // 73 // We allow the user to set a min, start, and max size for waitBuf. 74 // Whenever a sync starts, bytes are transferred from waitBuf into 75 // syncBuf and a timer is started. When a sync completes 76 // successfully, the number of bytes synced is added to the allowed 77 // size of waitBuf (i.e., "additive increase" == exponential growth). 78 // However, if the number of sync'd bytes is smaller than the min 79 // waitBuf size, we don't do additive increase (because we haven't 80 // really tested the throughput of the server connection in that case). 81 // 82 // If the sync takes more than 33% of half the overall operation 83 // timeout, the size of waitBuf is reduced by that same percentage 84 // (i.e., "multiplicative decrease"), and any outstanding bytes in the 85 // sync will not be used in the "additive increase" phase when the 86 // sync completes (they are considered "lost" in the TCP analogy, even 87 // though they should eventually succeed). The 33% limit was chosen 88 // mostly by trial and error, although if you assume that 89 // capacity(waitBuf) will double after each sync, then `2*len(syncBuf) 90 // == capacity(waitBuf)`, so at any given point there can be about 91 // 3*capacity(syncBuf) bytes buffered; so if syncBuf can't be sync'd 92 // in a third of the overall timeout, the next waitBuf should be 93 // reduced. 94 // 95 // Finally, we need to make sure that the Write calls that are filling 96 // up waitBuf while a sync is happening don't timeout. But, we also 97 // need to fill waitBuf quickly, so that the next sync is ready to go 98 // as soon as the first one completes. Here we implement a 99 // compromise. Allow waitBuf to fill up instantly until it holds 100 // capacity(syncBuf) bytes. After that, allow it to fill up to 101 // 2*capacity(syncBuf), but artificially delay each write by adding 102 // backpressure, by some fraction of the system operation timeout that 103 // matches the fraction of the progress the buffer has made between 104 // capacity(syncBuf) and 2*capacity(syncBuf). As soon as the sync 105 // completes, any delayed write is unblocked and gets to start filling 106 // up the buffers again. 107 // 108 // To avoid keeping the buffer capacity large when network conditions 109 // suddenly worsen (say after a laptop sleep when it comes back online 110 // on a new, worse network), the capacity is reset back to the minimum 111 // if resetBufferCapTime passes without any large syncs. TODO: in the 112 // future it might make sense to decrease the buffer capacity, rather 113 // than resetting it to the minimum? 114 type DirtyBlockCacheStandard struct { 115 clock idutil.Clock 116 log logger.Logger 117 vlog *libkb.VDebugLog 118 reqWg sync.WaitGroup 119 120 // requestsChan is a queue for channels that should be closed when 121 // permission is granted to dirty new data. 122 requestsChan chan dirtyReq 123 // bytesDecreasedChan is signalled when syncs have finished or dirty 124 // blocks have been deleted. 125 bytesDecreasedChan chan struct{} 126 // shutdownChan is closed when Shutdown is called. 127 shutdownChan chan struct{} 128 // blockedChanForTesting sends out the number of bytes of the 129 // request currently waiting. Sends out -1 when the request is 130 // accepted. Used only for testing. 131 blockedChanForTesting chan<- int64 132 133 // The minimum (and initial) capacity of the sync buffer. 134 minSyncBufCap int64 135 // The maximum capacity of the sync buffer. Also used as the 136 // denominator when calculating backpressure, such that the closer 137 // we are to reaching the maximum size (over and above the current 138 // sync buffer), the more write requests will be delayed. 139 maxSyncBufCap int64 140 141 // After how long without a syncBufferCap-sized sync will 142 // syncBufferCap be reset automatically back down to the minimum, 143 // to avoid keeping it too high as network conditions change? 144 resetBufferCapTime time.Duration 145 146 shutdownLock sync.RWMutex 147 isShutdown bool 148 149 lock sync.RWMutex 150 cache map[dirtyBlockID]Block 151 syncBufBytes int64 152 waitBufBytes int64 153 syncBufferCap int64 154 ignoreSyncBytes int64 // these bytes have "timed out" 155 syncStarted time.Time 156 resetter *time.Timer 157 } 158 159 // NewDirtyBlockCacheStandard constructs a new BlockCacheStandard 160 // instance. The min and max buffer capacities define the possible 161 // range of how many bytes we'll try to sync in any one sync, and the 162 // start size defines the initial buffer size. 163 func NewDirtyBlockCacheStandard( 164 clock idutil.Clock, log logger.Logger, vlog *libkb.VDebugLog, 165 minSyncBufCap int64, maxSyncBufCap int64, 166 startSyncBufCap int64) *DirtyBlockCacheStandard { 167 d := &DirtyBlockCacheStandard{ 168 clock: clock, 169 log: log, 170 vlog: vlog, 171 requestsChan: make(chan dirtyReq, 1000), 172 bytesDecreasedChan: make(chan struct{}, 1), 173 shutdownChan: make(chan struct{}), 174 cache: make(map[dirtyBlockID]Block), 175 minSyncBufCap: minSyncBufCap, 176 maxSyncBufCap: maxSyncBufCap, 177 syncBufferCap: startSyncBufCap, 178 resetBufferCapTime: resetBufferCapTimeDefault, 179 } 180 d.reqWg.Add(1) 181 go d.processPermission() 182 return d 183 } 184 185 // SimpleDirtyBlockCacheStandard that can only handle block 186 // put/get/delete requests; it cannot track dirty bytes. 187 func SimpleDirtyBlockCacheStandard() *DirtyBlockCacheStandard { 188 return &DirtyBlockCacheStandard{ 189 cache: make(map[dirtyBlockID]Block), 190 } 191 } 192 193 // Get implements the DirtyBlockCache interface for 194 // DirtyBlockCacheStandard. 195 func (d *DirtyBlockCacheStandard) Get( 196 _ context.Context, _ tlf.ID, ptr BlockPointer, branch BranchName) ( 197 Block, error) { 198 block := func() Block { 199 dirtyID := dirtyBlockID{ 200 id: ptr.ID, 201 refNonce: ptr.RefNonce, 202 branch: branch, 203 } 204 d.lock.RLock() 205 defer d.lock.RUnlock() 206 return d.cache[dirtyID] 207 }() 208 if block != nil { 209 return block, nil 210 } 211 212 return nil, NoSuchBlockError{ptr.ID} 213 } 214 215 // Put implements the DirtyBlockCache interface for 216 // DirtyBlockCacheStandard. 217 func (d *DirtyBlockCacheStandard) Put( 218 _ context.Context, _ tlf.ID, ptr BlockPointer, branch BranchName, 219 block Block) error { 220 dirtyID := dirtyBlockID{ 221 id: ptr.ID, 222 refNonce: ptr.RefNonce, 223 branch: branch, 224 } 225 226 d.lock.Lock() 227 defer d.lock.Unlock() 228 d.cache[dirtyID] = block 229 return nil 230 } 231 232 // Delete implements the DirtyBlockCache interface for 233 // DirtyBlockCacheStandard. 234 func (d *DirtyBlockCacheStandard) Delete(_ tlf.ID, ptr BlockPointer, 235 branch BranchName) error { 236 dirtyID := dirtyBlockID{ 237 id: ptr.ID, 238 refNonce: ptr.RefNonce, 239 branch: branch, 240 } 241 242 d.lock.Lock() 243 defer d.lock.Unlock() 244 delete(d.cache, dirtyID) 245 return nil 246 } 247 248 // IsDirty implements the DirtyBlockCache interface for 249 // DirtyBlockCacheStandard. 250 func (d *DirtyBlockCacheStandard) IsDirty(_ tlf.ID, ptr BlockPointer, 251 branch BranchName) (isDirty bool) { 252 dirtyID := dirtyBlockID{ 253 id: ptr.ID, 254 refNonce: ptr.RefNonce, 255 branch: branch, 256 } 257 258 d.lock.RLock() 259 defer d.lock.RUnlock() 260 _, isDirty = d.cache[dirtyID] 261 return 262 } 263 264 // IsAnyDirty implements the DirtyBlockCache interface for 265 // DirtyBlockCacheStandard. 266 func (d *DirtyBlockCacheStandard) IsAnyDirty(_ tlf.ID) bool { 267 d.lock.RLock() 268 defer d.lock.RUnlock() 269 return len(d.cache) > 0 || d.syncBufBytes > 0 || d.waitBufBytes > 0 270 } 271 272 const backpressureSlack = 1 * time.Second 273 274 // calcBackpressure returns how much longer a given request should be 275 // blocked, as a function of its deadline and how past full the 276 // syncing buffer is. In its lifetime, the request should be blocked 277 // by roughly the same fraction of its total deadline as how past full 278 // the buffer is. This will let KBFS slow down writes according to 279 // how slow the background Syncs are, so we don't accumulate more 280 // bytes to Sync than we can handle. See KBFS-731. 281 func (d *DirtyBlockCacheStandard) calcBackpressure(start time.Time, 282 deadline time.Time) time.Duration { 283 d.lock.RLock() 284 defer d.lock.RUnlock() 285 // We don't want to use the whole deadline, so cut it some slack. 286 totalReqTime := deadline.Sub(start) - backpressureSlack 287 if totalReqTime <= 0 { 288 return 0 289 } 290 291 // Keep the window full in preparation for the next sync, after 292 // it's full start applying backpressure. 293 if d.waitBufBytes < d.syncBufferCap { 294 return 0 295 } 296 297 // The backpressure is proportional to how far our overage is 298 // toward filling up our next sync buffer. 299 backpressureFrac := float64(d.waitBufBytes-d.syncBufferCap) / 300 float64(d.syncBufferCap) 301 if backpressureFrac > 1.0 { 302 backpressureFrac = 1.0 303 } 304 totalBackpressure := time.Duration( 305 float64(totalReqTime) * backpressureFrac) 306 timeSpentSoFar := d.clock.Now().Sub(start) 307 if totalBackpressure <= timeSpentSoFar { 308 return 0 309 } 310 311 // How much time do we have left, given how much time this request 312 // has waited so far? 313 return totalBackpressure - timeSpentSoFar 314 } 315 316 func (d *DirtyBlockCacheStandard) acceptNewWrite(newBytes int64) bool { 317 d.lock.Lock() 318 defer d.lock.Unlock() 319 // Accept any write, as long as we're not already over the limits. 320 // Allow the total dirty bytes to get close to double the max 321 // buffer size, to allow us to fill up the buffer for the next 322 // sync. 323 canAccept := d.waitBufBytes < d.maxSyncBufCap*2 324 if canAccept { 325 d.waitBufBytes += newBytes 326 } 327 328 return canAccept 329 } 330 331 func (d *DirtyBlockCacheStandard) maybeDecreaseBuffer(start time.Time, 332 deadline time.Time, soFar float64) (bool, time.Duration, float64) { 333 // Update syncBufferCap if the write has been blocked for more 334 // than half of its timeout. (We use half the timeout in case a 335 // user Sync operation, which can't be subjected to backpressure, 336 // is blocked by a background Sync operation when waitBuf is 337 // nearly full.) 338 allowedTimeout := float64(deadline.Sub(start)) / 2.0 339 timeoutUsed := d.clock.Now().Sub(start) 340 fracTimeoutUsed := float64(timeoutUsed) / allowedTimeout 341 if fracTimeoutUsed >= 0.33 { 342 d.lock.Lock() 343 defer d.lock.Unlock() 344 // Decrease the syncBufferCap by the percentage of the timeout 345 // we're using, minus the percentage we've already decreased 346 // it so far. TODO: a more logical algorithm would probably 347 // keep track of what the syncBufferCap was before the Sync 348 // started, and multiply that by the entire fracTimeoutUsed, 349 // since subtracting percentages in this way doesn't make a 350 // whole lot of sense. 351 d.syncBufferCap = int64(float64(d.syncBufferCap) * 352 (1 - (fracTimeoutUsed - soFar))) 353 if d.syncBufferCap < d.minSyncBufCap { 354 d.syncBufferCap = d.minSyncBufCap 355 } 356 d.log.CDebugf(context.TODO(), "Writes blocked for %s (%f%% of timeout), "+ 357 "syncBufferCap=%d", timeoutUsed, fracTimeoutUsed*100, 358 d.syncBufferCap) 359 if d.syncBufBytes > d.ignoreSyncBytes { 360 d.ignoreSyncBytes = d.syncBufBytes 361 } 362 return true, time.Duration(allowedTimeout), fracTimeoutUsed 363 } 364 365 // If we haven't decreased the buffer yet, make sure we get a 366 // wake-up call at the right time. 367 maxWakeup := allowedTimeout / 3.0 368 return false, time.Duration(maxWakeup) - timeoutUsed, soFar 369 } 370 371 func (d *DirtyBlockCacheStandard) getSyncStarted() time.Time { 372 d.lock.RLock() 373 defer d.lock.RUnlock() 374 return d.syncStarted 375 } 376 377 func (d *DirtyBlockCacheStandard) getSyncBufferCap() int64 { 378 d.lock.RLock() 379 defer d.lock.RUnlock() 380 return d.syncBufferCap 381 } 382 383 func (d *DirtyBlockCacheStandard) processPermission() { 384 defer d.reqWg.Done() 385 // Keep track of the most-recently seen request across loop 386 // iterations, because we aren't necessarily going to be able to 387 // deal with it as soon as we see it (since we might be past our 388 // limits already). 389 var currentReq dirtyReq 390 var backpressure time.Duration 391 var maxWakeup time.Duration 392 decreased := false 393 var fracDeadlineSoFar float64 394 var lastKnownTimeout time.Duration 395 for { 396 reqChan := d.requestsChan 397 if currentReq.respChan != nil { 398 // We are already waiting on a request, so don't bother 399 // trying to read another request from the requests chan. 400 reqChan = nil 401 402 // If we haven't decreased the buffer size yet, make sure 403 // we wake up in time to do that. 404 if !decreased && (backpressure <= 0 || maxWakeup < backpressure) { 405 backpressure = maxWakeup 406 } 407 } else if !d.getSyncStarted().IsZero() { 408 // There are no requests pending, but there is still a 409 // sync pending. 410 backpressure = maxWakeup 411 } 412 413 var bpTimer <-chan time.Time 414 if backpressure > 0 { 415 bpTimer = time.After(backpressure) 416 } 417 418 newReq := false 419 select { 420 case <-d.shutdownChan: 421 return 422 case <-d.bytesDecreasedChan: 423 case <-bpTimer: 424 case r := <-reqChan: 425 currentReq = r 426 newReq = true 427 decreased = false 428 } 429 430 if currentReq.respChan != nil || maxWakeup > 0 { 431 syncStarted := d.getSyncStarted() 432 // Has this sync been blocking so long that we should 433 // decrease the buffer size? 434 if !syncStarted.IsZero() { 435 deadline := syncStarted.Add(lastKnownTimeout) 436 decreased, maxWakeup, fracDeadlineSoFar = 437 d.maybeDecreaseBuffer(syncStarted, 438 deadline, fracDeadlineSoFar) 439 } else { 440 maxWakeup = 0 441 } 442 } 443 444 if currentReq.respChan != nil { 445 lastKnownTimeout = currentReq.deadline.Sub(currentReq.start) 446 // Apply any backpressure? 447 backpressure = d.calcBackpressure(currentReq.start, 448 currentReq.deadline) 449 switch { 450 case backpressure == 0 && d.acceptNewWrite(currentReq.bytes): 451 // If we have an active request, and we have room in 452 // our buffers to deal with it, grant permission to 453 // the requestor by closing the response channel. 454 close(currentReq.respChan) 455 currentReq = dirtyReq{} 456 if d.blockedChanForTesting != nil { 457 d.blockedChanForTesting <- -1 458 } 459 case d.blockedChanForTesting != nil && newReq: 460 // Otherwise, if this is the first time we've 461 // considered this request, inform any tests that the 462 // request is blocked. 463 d.blockedChanForTesting <- currentReq.bytes 464 case backpressure != 0: 465 func() { 466 d.lock.Lock() 467 defer d.lock.Unlock() 468 if d.syncStarted.IsZero() { 469 // TODO: in this case where there are multiple 470 // concurrent Syncs from multiple TLFs, this 471 // might not correctly capture the start time 472 // of the Nth Sync. We might want to assign 473 // each Sync its own unique ID somehow, so we 474 // can track them separately and more 475 // accurately. 476 d.syncStarted = d.clock.Now() 477 fracDeadlineSoFar = 0 478 } 479 d.log.CDebugf(context.TODO(), "Applying backpressure %s", backpressure) 480 }() 481 } 482 } 483 } 484 } 485 486 // RequestPermissionToDirty implements the DirtyBlockCache interface 487 // for DirtyBlockCacheStandard. 488 func (d *DirtyBlockCacheStandard) RequestPermissionToDirty( 489 ctx context.Context, _ tlf.ID, estimatedDirtyBytes int64) ( 490 DirtyPermChan, error) { 491 d.shutdownLock.RLock() 492 defer d.shutdownLock.RUnlock() 493 if d.isShutdown { 494 return nil, ShutdownHappenedError{} 495 } 496 497 if estimatedDirtyBytes < 0 { 498 panic("Must request permission for a non-negative number of bytes.") 499 } 500 c := make(chan struct{}) 501 502 // No need to wait to write 0 bytes. 503 if estimatedDirtyBytes == 0 { 504 close(c) 505 return c, nil 506 } 507 508 now := d.clock.Now() 509 deadline, ok := ctx.Deadline() 510 defaultDeadline := now.Add(BackgroundTaskTimeout / 2) 511 if !ok || deadline.After(defaultDeadline) { 512 // Use half of the background task timeout, to make sure we 513 // never get close to a timeout in a background task. 514 deadline = defaultDeadline 515 } 516 req := dirtyReq{c, estimatedDirtyBytes, now, deadline} 517 select { 518 case d.requestsChan <- req: 519 return c, nil 520 case <-ctx.Done(): 521 return nil, ctx.Err() 522 } 523 } 524 525 func (d *DirtyBlockCacheStandard) signalDecreasedBytes() { 526 select { 527 case d.bytesDecreasedChan <- struct{}{}: 528 default: 529 // Already something queued there, and one is enough. 530 } 531 } 532 533 func (d *DirtyBlockCacheStandard) updateWaitBufLocked(bytes int64) { 534 d.waitBufBytes += bytes 535 if d.waitBufBytes < 0 { 536 // It would be better if we didn't have this check, but it's 537 // hard for folderBlockOps to account correctly when bytes in 538 // a syncing block are overwritten, and then the write is 539 // deferred (see KBFS-2157). 540 d.waitBufBytes = 0 541 } 542 } 543 544 // UpdateUnsyncedBytes implements the DirtyBlockCache interface for 545 // DirtyBlockCacheStandard. 546 func (d *DirtyBlockCacheStandard) UpdateUnsyncedBytes(_ tlf.ID, 547 newUnsyncedBytes int64, wasSyncing bool) { 548 d.lock.Lock() 549 defer d.lock.Unlock() 550 if wasSyncing { 551 d.syncBufBytes += newUnsyncedBytes 552 } else { 553 d.updateWaitBufLocked(newUnsyncedBytes) 554 } 555 if newUnsyncedBytes < 0 { 556 d.signalDecreasedBytes() 557 } 558 } 559 560 // UpdateSyncingBytes implements the DirtyBlockCache interface for 561 // DirtyBlockCacheStandard. 562 func (d *DirtyBlockCacheStandard) UpdateSyncingBytes(_ tlf.ID, size int64) { 563 d.lock.Lock() 564 defer d.lock.Unlock() 565 d.syncBufBytes += size 566 d.updateWaitBufLocked(-size) 567 d.signalDecreasedBytes() 568 } 569 570 // BlockSyncFinished implements the DirtyBlockCache interface for 571 // DirtyBlockCacheStandard. 572 func (d *DirtyBlockCacheStandard) BlockSyncFinished(_ tlf.ID, size int64) { 573 d.lock.Lock() 574 defer d.lock.Unlock() 575 if size > 0 { 576 d.syncBufBytes -= size 577 } else { 578 // The block will be retried, so put it back on the waitBuf 579 d.updateWaitBufLocked(-size) 580 } 581 if size > 0 { 582 d.signalDecreasedBytes() 583 } 584 } 585 586 func (d *DirtyBlockCacheStandard) resetBufferCap() { 587 d.lock.Lock() 588 defer d.lock.Unlock() 589 d.log.CDebugf(context.TODO(), "Resetting syncBufferCap from %d to %d", d.syncBufferCap, 590 d.minSyncBufCap) 591 d.syncBufferCap = d.minSyncBufCap 592 d.resetter = nil 593 if d.blockedChanForTesting != nil { 594 d.blockedChanForTesting <- -1 595 } 596 } 597 598 // SyncFinished implements the DirtyBlockCache interface for 599 // DirtyBlockCacheStandard. 600 func (d *DirtyBlockCacheStandard) SyncFinished(_ tlf.ID, size int64) { 601 d.lock.Lock() 602 defer d.lock.Unlock() 603 if size <= 0 { 604 return 605 } 606 d.syncStarted = time.Time{} 607 608 // If the outstanding bytes have timed out, don't count them 609 // towards the buffer increase. 610 ignore := d.ignoreSyncBytes 611 if ignore > size { 612 ignore = size 613 } 614 bufferIncrease := size - ignore 615 d.ignoreSyncBytes -= ignore 616 617 // If the sync was a reasonably large fraction of the current 618 // buffer capacity, restart the reset timer. 619 if size >= d.syncBufferCap/2 { 620 if d.resetter != nil { 621 d.resetter.Stop() 622 } 623 d.resetter = time.AfterFunc(d.resetBufferCapTime, d.resetBufferCap) 624 } 625 626 // Only increase the buffer size if we sent over a lot of bytes. 627 // We don't want a series of small writes to increase the buffer 628 // size, since that doesn't give us any real information about the 629 // throughput of the connection. 630 if bufferIncrease >= d.syncBufferCap { 631 d.syncBufferCap += bufferIncrease 632 if d.syncBufferCap > d.maxSyncBufCap { 633 d.syncBufferCap = d.maxSyncBufCap 634 } 635 } 636 d.signalDecreasedBytes() 637 d.vlog.CLogf( 638 context.TODO(), libkb.VLog1, 639 "Finished syncing %d bytes, syncBufferCap=%d, waitBuf=%d, ignored=%d", 640 size, d.syncBufferCap, d.waitBufBytes, ignore) 641 } 642 643 // ShouldForceSync implements the DirtyBlockCache interface for 644 // DirtyBlockCacheStandard. 645 func (d *DirtyBlockCacheStandard) ShouldForceSync(_ tlf.ID) bool { 646 d.lock.RLock() 647 defer d.lock.RUnlock() 648 // TODO: Fill up to likely block boundaries? 649 return d.waitBufBytes >= d.syncBufferCap 650 } 651 652 // Size returns the number of blocks currently in the cache. 653 func (d *DirtyBlockCacheStandard) Size() int { 654 d.lock.RLock() 655 defer d.lock.RUnlock() 656 return len(d.cache) 657 } 658 659 // Shutdown implements the DirtyBlockCache interface for 660 // DirtyBlockCacheStandard. 661 func (d *DirtyBlockCacheStandard) Shutdown() error { 662 func() { 663 d.shutdownLock.Lock() 664 defer d.shutdownLock.Unlock() 665 d.isShutdown = true 666 close(d.shutdownChan) 667 }() 668 669 d.reqWg.Wait() 670 close(d.requestsChan) 671 d.lock.Lock() 672 defer d.lock.Unlock() 673 // Clear out the remaining requests 674 for req := range d.requestsChan { 675 d.updateWaitBufLocked(req.bytes) 676 } 677 if d.syncBufBytes != 0 || d.waitBufBytes != 0 || d.ignoreSyncBytes != 0 { 678 return fmt.Errorf("Unexpected dirty bytes leftover on shutdown: "+ 679 "syncBuf=%d, waitBuf=%d, ignore=%d", 680 d.syncBufBytes, d.waitBufBytes, d.ignoreSyncBytes) 681 } 682 return nil 683 }