github.com/keybase/client/go@v0.0.0-20241007131713-f10651d043c8/kbfs/libkbfs/disk_block_cache.go (about) 1 // Copyright 2017 Keybase Inc. All rights reserved. 2 // Use of this source code is governed by a BSD 3 // license that can be found in the LICENSE file. 4 5 package libkbfs 6 7 import ( 8 "fmt" 9 "io" 10 "math" 11 "math/rand" 12 "path/filepath" 13 "sort" 14 "sync" 15 "time" 16 17 "github.com/keybase/client/go/kbfs/data" 18 "github.com/keybase/client/go/kbfs/kbfsblock" 19 "github.com/keybase/client/go/kbfs/kbfscrypto" 20 "github.com/keybase/client/go/kbfs/kbfshash" 21 "github.com/keybase/client/go/kbfs/kbfsmd" 22 "github.com/keybase/client/go/kbfs/ldbutils" 23 "github.com/keybase/client/go/kbfs/tlf" 24 "github.com/keybase/client/go/logger" 25 "github.com/keybase/client/go/protocol/keybase1" 26 "github.com/pkg/errors" 27 "github.com/syndtr/goleveldb/leveldb" 28 "github.com/syndtr/goleveldb/leveldb/filter" 29 "github.com/syndtr/goleveldb/leveldb/opt" 30 "github.com/syndtr/goleveldb/leveldb/storage" 31 "github.com/syndtr/goleveldb/leveldb/util" 32 "golang.org/x/net/context" 33 ) 34 35 const ( 36 defaultBlockCacheTableSize int = 50 * opt.MiB 37 defaultBlockCacheBlockSize int = 4 * opt.MiB 38 defaultBlockCacheCapacity int = 8 * opt.MiB 39 evictionConsiderationFactor int = 3 40 minNumBlocksToEvictInBatch int = 10 41 maxNumBlocksToEvictInBatch int = 500 42 defaultNumBlocksToEvictOnClear int = 100 43 defaultNumUnmarkedBlocksToCheck int = 100 44 defaultClearTickerDuration = 1 * time.Second 45 maxEvictionsPerPut int = 100 46 blockDbFilename string = "diskCacheBlocks.leveldb" 47 metaDbFilename string = "diskCacheMetadata.leveldb" 48 tlfDbFilename string = "diskCacheTLF.leveldb" 49 lastUnrefDbFilename string = "diskCacheLastUnref.leveldb" 50 initialDiskBlockCacheVersion uint64 = 1 51 currentDiskBlockCacheVersion uint64 = initialDiskBlockCacheVersion 52 syncCacheName string = "SyncBlockCache" 53 workingSetCacheName string = "WorkingSetBlockCache" 54 crDirtyBlockCacheName string = "DirtyBlockCache" 55 minDiskBlockWriteBufferSize = 3 * data.MaxBlockSizeBytesDefault // ~ 1 MB 56 deleteCompactThreshold int = 25 57 compactTimer = time.Minute * 5 58 ) 59 60 var errTeamOrUnknownTLFAddedAsHome = errors.New( 61 "Team or Unknown TLF added to disk block cache as home TLF") 62 63 type evictionPriority int 64 65 const ( 66 priorityNotHome evictionPriority = iota 67 priorityPublicHome 68 priorityPrivateHome 69 ) 70 71 // DiskBlockCacheLocal is the standard implementation for DiskBlockCache. 72 type DiskBlockCacheLocal struct { 73 config diskBlockCacheConfig 74 log logger.Logger 75 maxBlockID []byte 76 dirPath string 77 78 clearTickerDuration time.Duration 79 numBlocksToEvictOnClear int 80 numUnmarkedBlocksToCheck int 81 82 // Track the cache hit rate and eviction rate 83 hitMeter *ldbutils.CountMeter 84 missMeter *ldbutils.CountMeter 85 putMeter *ldbutils.CountMeter 86 updateMeter *ldbutils.CountMeter 87 evictCountMeter *ldbutils.CountMeter 88 evictSizeMeter *ldbutils.CountMeter 89 deleteCountMeter *ldbutils.CountMeter 90 deleteSizeMeter *ldbutils.CountMeter 91 92 // Protect the disk caches from being shutdown while they're being 93 // accessed, and mutable data. 94 lock sync.RWMutex 95 blockDb *ldbutils.LevelDb 96 metaDb *ldbutils.LevelDb 97 tlfDb *ldbutils.LevelDb 98 lastUnrefDb *ldbutils.LevelDb 99 cacheType diskLimitTrackerType 100 // Track the number of blocks in the cache per TLF and overall. 101 tlfCounts map[tlf.ID]int 102 numBlocks int 103 // Track the number of blocks in the cahce per eviction priority, 104 // for easy eviction counting. 105 priorityBlockCounts map[evictionPriority]int 106 priorityTlfMap map[evictionPriority]map[tlf.ID]int 107 // Track the aggregate size of blocks in the cache per TLF and overall. 108 tlfSizes map[tlf.ID]uint64 109 // Track the last unref'd revisions for each TLF. 110 tlfLastUnrefs map[tlf.ID]kbfsmd.Revision 111 // Don't evict files from the user's private or public home directory. 112 // Higher numbers are more important not to evict. 113 homeDirs map[tlf.ID]evictionPriority 114 115 // currBytes gets its own lock, since tests need to access it 116 // directly and taking the full lock causes deadlocks under some 117 // situations. 118 currBytesLock sync.RWMutex 119 currBytes uint64 120 121 compactCh chan struct{} 122 useCh chan struct{} 123 startedCh chan struct{} 124 startErrCh chan struct{} 125 shutdownCh chan struct{} 126 doneCh chan struct{} 127 128 closer func() 129 } 130 131 // DiskBlockCacheStartState represents whether this disk block cache has 132 // started or failed. 133 type DiskBlockCacheStartState int 134 135 // String allows DiskBlockCacheStartState to be output as a string. 136 func (s DiskBlockCacheStartState) String() string { 137 switch s { 138 case DiskBlockCacheStartStateStarting: 139 return "starting" 140 case DiskBlockCacheStartStateStarted: 141 return "started" 142 case DiskBlockCacheStartStateFailed: 143 return "failed" 144 default: 145 return "unknown" 146 } 147 } 148 149 const ( 150 // DiskBlockCacheStartStateStarting represents when the cache is starting. 151 DiskBlockCacheStartStateStarting DiskBlockCacheStartState = iota 152 // DiskBlockCacheStartStateStarted represents when the cache has started. 153 DiskBlockCacheStartStateStarted 154 // DiskBlockCacheStartStateFailed represents when the cache has failed to 155 // start. 156 DiskBlockCacheStartStateFailed 157 ) 158 159 // DiskBlockCacheStatus represents the status of the disk cache. 160 type DiskBlockCacheStatus struct { 161 StartState DiskBlockCacheStartState 162 NumBlocks uint64 163 BlockBytes uint64 164 CurrByteLimit uint64 165 LastUnrefCount uint64 166 Hits ldbutils.MeterStatus 167 Misses ldbutils.MeterStatus 168 Puts ldbutils.MeterStatus 169 MetadataUpdates ldbutils.MeterStatus 170 NumEvicted ldbutils.MeterStatus 171 SizeEvicted ldbutils.MeterStatus 172 NumDeleted ldbutils.MeterStatus 173 SizeDeleted ldbutils.MeterStatus 174 175 LocalDiskBytesAvailable uint64 176 LocalDiskBytesTotal uint64 177 178 BlockDBStats []string `json:",omitempty"` 179 MetaDBStats []string `json:",omitempty"` 180 TLFDBStats []string `json:",omitempty"` 181 LastUnrefStats []string `json:",omitempty"` 182 MemCompActive bool `json:",omitempty"` 183 TableCompActive bool `json:",omitempty"` 184 MetaMemCompActive bool `json:",omitempty"` 185 MetaTableCompActive bool `json:",omitempty"` 186 } 187 188 type lastUnrefEntry struct { 189 Rev kbfsmd.Revision 190 Ctime time.Time // Not used yet, but save it in case we ever need it. 191 } 192 193 // newDiskBlockCacheLocalFromStorage creates a new *DiskBlockCacheLocal 194 // with the passed-in storage.Storage interfaces as storage layers for each 195 // cache. 196 func newDiskBlockCacheLocalFromStorage( 197 config diskBlockCacheConfig, cacheType diskLimitTrackerType, 198 blockStorage, metadataStorage, tlfStorage, 199 lastUnrefStorage storage.Storage, mode InitMode) ( 200 cache *DiskBlockCacheLocal, err error) { 201 log := config.MakeLogger("KBC") 202 closers := make([]io.Closer, 0, 3) 203 closer := func() { 204 for _, c := range closers { 205 closeErr := c.Close() 206 if closeErr != nil { 207 log.Warning("Error closing leveldb or storage: %+v", closeErr) 208 } 209 } 210 } 211 defer func() { 212 if err != nil { 213 err = errors.WithStack(err) 214 closer() 215 } 216 }() 217 blockDbOptions := ldbutils.LeveldbOptions(mode) 218 blockDbOptions.CompactionTableSize = defaultBlockCacheTableSize 219 blockDbOptions.BlockSize = defaultBlockCacheBlockSize 220 blockDbOptions.BlockCacheCapacity = defaultBlockCacheCapacity 221 blockDbOptions.Filter = filter.NewBloomFilter(16) 222 if blockDbOptions.WriteBuffer < minDiskBlockWriteBufferSize { 223 blockDbOptions.WriteBuffer = minDiskBlockWriteBufferSize 224 } 225 blockDb, err := ldbutils.OpenLevelDbWithOptions(blockStorage, blockDbOptions) 226 if err != nil { 227 return nil, err 228 } 229 closers = append(closers, blockDb) 230 231 metaDb, err := ldbutils.OpenLevelDb(metadataStorage, mode) 232 if err != nil { 233 return nil, err 234 } 235 closers = append(closers, metaDb) 236 237 tlfDb, err := ldbutils.OpenLevelDb(tlfStorage, mode) 238 if err != nil { 239 return nil, err 240 } 241 closers = append(closers, tlfDb) 242 243 lastUnrefDb, err := ldbutils.OpenLevelDb(lastUnrefStorage, mode) 244 if err != nil { 245 return nil, err 246 } 247 closers = append(closers, lastUnrefDb) 248 249 maxBlockID, err := kbfshash.HashFromRaw( 250 kbfshash.MaxHashType, kbfshash.MaxDefaultHash[:]) 251 if err != nil { 252 return nil, err 253 } 254 startedCh := make(chan struct{}) 255 startErrCh := make(chan struct{}) 256 cache = &DiskBlockCacheLocal{ 257 config: config, 258 maxBlockID: maxBlockID.Bytes(), 259 clearTickerDuration: defaultClearTickerDuration, 260 numBlocksToEvictOnClear: defaultNumBlocksToEvictOnClear, 261 numUnmarkedBlocksToCheck: defaultNumUnmarkedBlocksToCheck, 262 cacheType: cacheType, 263 hitMeter: ldbutils.NewCountMeter(), 264 missMeter: ldbutils.NewCountMeter(), 265 putMeter: ldbutils.NewCountMeter(), 266 updateMeter: ldbutils.NewCountMeter(), 267 evictCountMeter: ldbutils.NewCountMeter(), 268 evictSizeMeter: ldbutils.NewCountMeter(), 269 deleteCountMeter: ldbutils.NewCountMeter(), 270 deleteSizeMeter: ldbutils.NewCountMeter(), 271 homeDirs: map[tlf.ID]evictionPriority{}, 272 log: log, 273 blockDb: blockDb, 274 metaDb: metaDb, 275 tlfDb: tlfDb, 276 lastUnrefDb: lastUnrefDb, 277 tlfCounts: map[tlf.ID]int{}, 278 priorityBlockCounts: map[evictionPriority]int{}, 279 priorityTlfMap: map[evictionPriority]map[tlf.ID]int{ 280 priorityPublicHome: {}, 281 priorityPrivateHome: {}, 282 priorityNotHome: {}, 283 }, 284 tlfSizes: map[tlf.ID]uint64{}, 285 tlfLastUnrefs: map[tlf.ID]kbfsmd.Revision{}, 286 compactCh: make(chan struct{}, 1), 287 useCh: make(chan struct{}, 1), 288 startedCh: startedCh, 289 startErrCh: startErrCh, 290 shutdownCh: make(chan struct{}), 291 doneCh: make(chan struct{}), 292 closer: closer, 293 } 294 // Sync the block counts asynchronously so syncing doesn't block init. 295 // Since this method blocks, any Get or Put requests to the disk block 296 // cache will block until this is done. The log will contain the beginning 297 // and end of this sync. 298 go func() { 299 err := cache.syncBlockCountsAndUnrefsFromDb() 300 if err != nil { 301 close(startErrCh) 302 closer() 303 log.Warning("Disabling disk block cache due to error syncing the "+ 304 "block counts from DB: %+v", err) 305 return 306 } 307 diskLimiter := cache.config.DiskLimiter() 308 if diskLimiter != nil && cache.useLimiter() { 309 // Notify the disk limiter of the disk cache's size once we've 310 // determined it. 311 ctx := context.Background() 312 cache.config.DiskLimiter().onSimpleByteTrackerEnable(ctx, 313 cache.cacheType, int64(cache.getCurrBytes())) 314 } 315 close(startedCh) 316 }() 317 318 // Only do background compaction on desktop for now, because on 319 // mobile we'd probably cause issues if we try to do it while 320 // backgrounded. 321 if mode.DiskCacheCompactionEnabled() { 322 go cache.compactLoop() 323 } 324 325 return cache, nil 326 } 327 328 // newDiskBlockCacheLocal creates a new *DiskBlockCacheLocal with a 329 // specified directory on the filesystem as storage. 330 func newDiskBlockCacheLocal(config diskBlockCacheConfig, 331 cacheType diskLimitTrackerType, dirPath string, mode InitMode) ( 332 cache *DiskBlockCacheLocal, err error) { 333 log := config.MakeLogger("DBC") 334 defer func() { 335 if err != nil { 336 log.Error("Error initializing disk cache: %+v", err) 337 config.GetPerfLog().CDebugf( 338 context.TODO(), 339 "KBFS couldn't initialize disk cache of type %s: %v", 340 cacheType, err) 341 } 342 }() 343 versionPath, err := ldbutils.GetVersionedPathForDb( 344 log, dirPath, "disk block cache", currentDiskBlockCacheVersion) 345 if err != nil { 346 return nil, err 347 } 348 blockDbPath := filepath.Join(versionPath, blockDbFilename) 349 blockStorage, err := storage.OpenFile(blockDbPath, false) 350 if err != nil { 351 return nil, err 352 } 353 defer func() { 354 if err != nil { 355 blockStorage.Close() 356 } 357 }() 358 metaDbPath := filepath.Join(versionPath, metaDbFilename) 359 metadataStorage, err := storage.OpenFile(metaDbPath, false) 360 if err != nil { 361 return nil, err 362 } 363 defer func() { 364 if err != nil { 365 metadataStorage.Close() 366 } 367 }() 368 tlfDbPath := filepath.Join(versionPath, tlfDbFilename) 369 tlfStorage, err := storage.OpenFile(tlfDbPath, false) 370 if err != nil { 371 return nil, err 372 } 373 defer func() { 374 if err != nil { 375 tlfStorage.Close() 376 } 377 }() 378 lastUnrefDbPath := filepath.Join(versionPath, lastUnrefDbFilename) 379 lastUnrefStorage, err := storage.OpenFile(lastUnrefDbPath, false) 380 if err != nil { 381 return nil, err 382 } 383 defer func() { 384 if err != nil { 385 lastUnrefStorage.Close() 386 } 387 }() 388 cache, err = newDiskBlockCacheLocalFromStorage(config, cacheType, 389 blockStorage, metadataStorage, tlfStorage, lastUnrefStorage, mode) 390 if err != nil { 391 return nil, err 392 } 393 cache.dirPath = dirPath 394 return cache, nil 395 } 396 397 func newDiskBlockCacheLocalForTest(config diskBlockCacheConfig, 398 cacheType diskLimitTrackerType) (*DiskBlockCacheLocal, error) { 399 return newDiskBlockCacheLocalFromStorage( 400 config, cacheType, storage.NewMemStorage(), 401 storage.NewMemStorage(), storage.NewMemStorage(), 402 storage.NewMemStorage(), &modeTest{modeDefault{}}) 403 } 404 405 func (cache *DiskBlockCacheLocal) useLimiter() bool { 406 return cache.cacheType != crDirtyBlockCacheLimitTrackerType 407 } 408 409 func (cache *DiskBlockCacheLocal) getCurrBytes() uint64 { 410 cache.currBytesLock.RLock() 411 defer cache.currBytesLock.RUnlock() 412 return cache.currBytes 413 } 414 415 func (cache *DiskBlockCacheLocal) setCurrBytes(b uint64) { 416 cache.currBytesLock.Lock() 417 defer cache.currBytesLock.Unlock() 418 cache.currBytes = b 419 } 420 421 func (cache *DiskBlockCacheLocal) addCurrBytes(b uint64) { 422 cache.currBytesLock.Lock() 423 defer cache.currBytesLock.Unlock() 424 cache.currBytes += b 425 } 426 427 func (cache *DiskBlockCacheLocal) subCurrBytes(b uint64) { 428 cache.currBytesLock.Lock() 429 defer cache.currBytesLock.Unlock() 430 if b <= cache.currBytes { 431 cache.currBytes -= b 432 } 433 } 434 435 // WaitUntilStarted waits until this cache has started. 436 func (cache *DiskBlockCacheLocal) WaitUntilStarted() error { 437 select { 438 case <-cache.startedCh: 439 return nil 440 case <-cache.startErrCh: 441 return DiskBlockCacheError{"error starting channel"} 442 } 443 } 444 445 func (cache *DiskBlockCacheLocal) decodeLastUnref(buf []byte) ( 446 rev kbfsmd.Revision, err error) { 447 var entry lastUnrefEntry 448 err = cache.config.Codec().Decode(buf, &entry) 449 if err != nil { 450 return kbfsmd.RevisionUninitialized, err 451 } 452 return entry.Rev, nil 453 } 454 455 func (cache *DiskBlockCacheLocal) encodeLastUnref(rev kbfsmd.Revision) ( 456 []byte, error) { 457 entry := lastUnrefEntry{ 458 Rev: rev, 459 Ctime: cache.config.Clock().Now(), 460 } 461 return cache.config.Codec().Encode(&entry) 462 } 463 464 func (cache *DiskBlockCacheLocal) syncBlockCountsAndUnrefsFromDb() error { 465 cache.log.Debug("+ syncBlockCountsAndUnrefsFromDb begin") 466 defer cache.log.Debug("- syncBlockCountsAndUnrefsFromDb end") 467 // We take a write lock for this to prevent any reads from happening while 468 // we're syncing the block counts. 469 cache.lock.Lock() 470 defer cache.lock.Unlock() 471 472 tlfCounts := make(map[tlf.ID]int) 473 tlfSizes := make(map[tlf.ID]uint64) 474 priorityBlockCounts := make(map[evictionPriority]int) 475 priorityTlfMap := map[evictionPriority]map[tlf.ID]int{ 476 priorityNotHome: {}, 477 priorityPublicHome: {}, 478 priorityPrivateHome: {}, 479 } 480 numBlocks := 0 481 totalSize := uint64(0) 482 iter := cache.metaDb.NewIterator(nil, nil) 483 defer iter.Release() 484 for iter.Next() { 485 metadata := DiskBlockCacheMetadata{} 486 err := cache.config.Codec().Decode(iter.Value(), &metadata) 487 if err != nil { 488 return err 489 } 490 size := uint64(metadata.BlockSize) 491 tlfCounts[metadata.TlfID]++ 492 tlfSizes[metadata.TlfID] += size 493 priorityBlockCounts[cache.homeDirs[metadata.TlfID]]++ 494 priorityTlfMap[cache.homeDirs[metadata.TlfID]][metadata.TlfID]++ 495 numBlocks++ 496 totalSize += size 497 } 498 cache.tlfCounts = tlfCounts 499 cache.numBlocks = numBlocks 500 cache.tlfSizes = tlfSizes 501 cache.setCurrBytes(totalSize) 502 cache.priorityTlfMap = priorityTlfMap 503 cache.priorityBlockCounts = priorityBlockCounts 504 505 cache.log.Debug("| syncBlockCountsAndUnrefsFromDb block counts done") 506 507 tlfLastUnrefs := make(map[tlf.ID]kbfsmd.Revision) 508 lastUnrefIter := cache.lastUnrefDb.NewIterator(nil, nil) 509 defer lastUnrefIter.Release() 510 for lastUnrefIter.Next() { 511 var tlfID tlf.ID 512 err := tlfID.UnmarshalBinary(lastUnrefIter.Key()) 513 if err != nil { 514 return err 515 } 516 517 rev, err := cache.decodeLastUnref(lastUnrefIter.Value()) 518 if err != nil { 519 return err 520 } 521 tlfLastUnrefs[tlfID] = rev 522 } 523 cache.tlfLastUnrefs = tlfLastUnrefs 524 525 return nil 526 } 527 528 // tlfKey generates a TLF cache key from a tlf.ID and a binary-encoded block 529 // ID. 530 func (*DiskBlockCacheLocal) tlfKey(tlfID tlf.ID, blockKey []byte) []byte { 531 return append(tlfID.Bytes(), blockKey...) 532 } 533 534 // updateMetadataLocked updates the LRU time of a block in the LRU cache to 535 // the current time. 536 func (cache *DiskBlockCacheLocal) updateMetadataLocked(ctx context.Context, 537 blockKey []byte, metadata DiskBlockCacheMetadata, metered bool) error { 538 metadata.LRUTime.Time = cache.config.Clock().Now() 539 encodedMetadata, err := cache.config.Codec().Encode(&metadata) 540 if err != nil { 541 return err 542 } 543 var putMeter *ldbutils.CountMeter 544 if ldbutils.Metered { 545 putMeter = cache.updateMeter 546 } 547 err = cache.metaDb.PutWithMeter(blockKey, encodedMetadata, putMeter) 548 if err != nil { 549 cache.log.CWarningf(ctx, "Error writing to disk cache meta "+ 550 "database: %+v", err) 551 } 552 return err 553 } 554 555 // getMetadataLocked retrieves the metadata for a block in the cache, or 556 // returns leveldb.ErrNotFound and a zero-valued metadata otherwise. 557 func (cache *DiskBlockCacheLocal) getMetadataLocked( 558 blockID kbfsblock.ID, metered bool) ( 559 metadata DiskBlockCacheMetadata, err error) { 560 var hitMeter, missMeter *ldbutils.CountMeter 561 if ldbutils.Metered { 562 hitMeter = cache.hitMeter 563 missMeter = cache.missMeter 564 } 565 566 metadataBytes, err := cache.metaDb.GetWithMeter( 567 blockID.Bytes(), hitMeter, missMeter) 568 if err != nil { 569 return DiskBlockCacheMetadata{}, err 570 } 571 err = cache.config.Codec().Decode(metadataBytes, &metadata) 572 return metadata, err 573 } 574 575 // getLRULocked retrieves the LRU time for a block in the cache, or returns 576 // leveldb.ErrNotFound and a zero-valued time.Time otherwise. 577 func (cache *DiskBlockCacheLocal) getLRULocked(blockID kbfsblock.ID) ( 578 time.Time, error) { 579 metadata, err := cache.getMetadataLocked(blockID, false) 580 if err != nil { 581 return time.Time{}, err 582 } 583 return metadata.LRUTime.Time, nil 584 } 585 586 // decodeBlockCacheEntry decodes a disk block cache entry buffer into an 587 // encoded block and server half. 588 func (cache *DiskBlockCacheLocal) decodeBlockCacheEntry(buf []byte) ([]byte, 589 kbfscrypto.BlockCryptKeyServerHalf, error) { 590 entry := diskBlockCacheEntry{} 591 err := cache.config.Codec().Decode(buf, &entry) 592 if err != nil { 593 return nil, kbfscrypto.BlockCryptKeyServerHalf{}, err 594 } 595 return entry.Buf, entry.ServerHalf, nil 596 } 597 598 // encodeBlockCacheEntry encodes an encoded block and serverHalf into a single 599 // buffer. 600 func (cache *DiskBlockCacheLocal) encodeBlockCacheEntry(buf []byte, 601 serverHalf kbfscrypto.BlockCryptKeyServerHalf) ([]byte, error) { 602 entry := diskBlockCacheEntry{ 603 Buf: buf, 604 ServerHalf: serverHalf, 605 } 606 return cache.config.Codec().Encode(&entry) 607 } 608 609 func (cache *DiskBlockCacheLocal) used() { 610 select { 611 case cache.useCh <- struct{}{}: 612 default: 613 } 614 } 615 616 // checkAndLockCache checks whether the cache is started. 617 func (cache *DiskBlockCacheLocal) checkCacheLocked(method string) (err error) { 618 defer func() { 619 if err == nil { 620 cache.used() 621 } 622 }() 623 624 select { 625 case <-cache.startedCh: 626 case <-cache.startErrCh: 627 // The cache will never be started. No need for a stack here since this 628 // could happen anywhere. 629 return DiskCacheStartingError{method} 630 default: 631 // If the cache hasn't started yet, return an error. No need for a 632 // stack here since this could happen anywhere. 633 return DiskCacheStartingError{method} 634 } 635 // shutdownCh has to be checked under lock, otherwise we can race. 636 select { 637 case <-cache.shutdownCh: 638 return errors.WithStack(DiskCacheClosedError{method}) 639 default: 640 } 641 if cache.blockDb == nil { 642 return errors.WithStack(DiskCacheClosedError{method}) 643 } 644 return nil 645 } 646 647 // Get implements the DiskBlockCache interface for DiskBlockCacheLocal. 648 func (cache *DiskBlockCacheLocal) Get( 649 ctx context.Context, tlfID tlf.ID, blockID kbfsblock.ID) (buf []byte, 650 serverHalf kbfscrypto.BlockCryptKeyServerHalf, 651 prefetchStatus PrefetchStatus, err error) { 652 cache.lock.RLock() 653 defer cache.lock.RUnlock() 654 err = cache.checkCacheLocked("Block(Get)") 655 if err != nil { 656 return nil, kbfscrypto.BlockCryptKeyServerHalf{}, NoPrefetch, err 657 } 658 659 blockKey := blockID.Bytes() 660 entry, err := cache.blockDb.Get(blockKey, nil) 661 if err != nil { 662 return nil, kbfscrypto.BlockCryptKeyServerHalf{}, NoPrefetch, 663 data.NoSuchBlockError{ID: blockID} 664 } 665 md, err := cache.getMetadataLocked(blockID, true) 666 if err != nil { 667 return nil, kbfscrypto.BlockCryptKeyServerHalf{}, NoPrefetch, err 668 } 669 err = cache.updateMetadataLocked(ctx, blockKey, md, ldbutils.Unmetered) 670 if err != nil { 671 return nil, kbfscrypto.BlockCryptKeyServerHalf{}, NoPrefetch, err 672 } 673 buf, serverHalf, err = cache.decodeBlockCacheEntry(entry) 674 return buf, serverHalf, md.PrefetchStatus(), err 675 } 676 677 // numBlocksToEvict estimates the number of blocks to evict to make 678 // enough room for new blocks, based on the average block size in the 679 // cache. 680 func (cache *DiskBlockCacheLocal) numBlocksToEvictLocked( 681 bytesAvailable int64) int { 682 if cache.numBlocks <= 0 || bytesAvailable > 0 { 683 return minNumBlocksToEvictInBatch 684 } 685 686 bytesPerBlock := int(cache.getCurrBytes()) / cache.numBlocks 687 toEvict := -int(bytesAvailable) / bytesPerBlock 688 if toEvict < minNumBlocksToEvictInBatch { 689 return minNumBlocksToEvictInBatch 690 } else if toEvict > maxNumBlocksToEvictInBatch { 691 return maxNumBlocksToEvictInBatch 692 } 693 return toEvict 694 } 695 696 func (cache *DiskBlockCacheLocal) evictUntilBytesAvailableLocked( 697 ctx context.Context, encodedLen int64) (hasEnoughSpace bool, err error) { 698 if !cache.useLimiter() { 699 return true, nil 700 } 701 for i := 0; i < maxEvictionsPerPut; i++ { 702 select { 703 // Ensure we don't loop infinitely 704 case <-ctx.Done(): 705 return false, ctx.Err() 706 default: 707 } 708 bytesAvailable, err := cache.config.DiskLimiter().reserveBytes( 709 ctx, cache.cacheType, encodedLen) 710 if err != nil { 711 cache.log.CWarningf(ctx, "Error obtaining space for the disk "+ 712 "block cache: %+v", err) 713 return false, err 714 } 715 if bytesAvailable >= 0 { 716 return true, nil 717 } 718 cache.log.CDebugf(ctx, "Need more bytes. Available: %d", bytesAvailable) 719 numRemoved, _, err := cache.evictLocked( 720 ctx, cache.numBlocksToEvictLocked(bytesAvailable)) 721 if err != nil { 722 return false, err 723 } 724 if numRemoved == 0 { 725 return false, errors.New("couldn't evict any more blocks from " + 726 "the disk block cache") 727 } 728 } 729 return false, nil 730 } 731 732 // Put implements the DiskBlockCache interface for DiskBlockCacheLocal. 733 func (cache *DiskBlockCacheLocal) Put( 734 ctx context.Context, tlfID tlf.ID, blockID kbfsblock.ID, buf []byte, 735 serverHalf kbfscrypto.BlockCryptKeyServerHalf) (err error) { 736 cache.lock.Lock() 737 defer cache.lock.Unlock() 738 err = cache.checkCacheLocked("Block(Put)") 739 if err != nil { 740 return err 741 } 742 743 blockLen := len(buf) 744 entry, err := cache.encodeBlockCacheEntry(buf, serverHalf) 745 if err != nil { 746 return err 747 } 748 encodedLen := int64(len(entry)) 749 defer func() { 750 cache.log.CDebugf(ctx, "Cache Put id=%s tlf=%s bSize=%d entrySize=%d "+ 751 "cacheType=%s err=%+v", blockID, tlfID, blockLen, encodedLen, 752 cache.cacheType, err) 753 }() 754 blockKey := blockID.Bytes() 755 hasKey, err := cache.blockDb.Has(blockKey, nil) 756 if err != nil { 757 cache.log.CDebugf(ctx, "Cache Put failed due to error from "+ 758 "blockDb.Has: %+v", err) 759 return err 760 } 761 if !hasKey { 762 if cache.cacheType == syncCacheLimitTrackerType { 763 bytesAvailable, err := cache.config.DiskLimiter().reserveBytes( 764 ctx, cache.cacheType, encodedLen) 765 if err != nil { 766 cache.log.CWarningf(ctx, "Error obtaining space for the disk "+ 767 "block cache: %+v", err) 768 return err 769 } 770 if bytesAvailable < 0 { 771 return data.CachePutCacheFullError{BlockID: blockID} 772 } 773 } else { 774 hasEnoughSpace, err := cache.evictUntilBytesAvailableLocked( 775 ctx, encodedLen) 776 if err != nil { 777 return err 778 } 779 if !hasEnoughSpace { 780 return data.CachePutCacheFullError{BlockID: blockID} 781 } 782 } 783 err = cache.blockDb.PutWithMeter(blockKey, entry, cache.putMeter) 784 if err != nil { 785 if cache.useLimiter() { 786 cache.config.DiskLimiter().commitOrRollback(ctx, 787 cache.cacheType, encodedLen, 0, false, "") 788 } 789 return err 790 } 791 if cache.useLimiter() { 792 cache.config.DiskLimiter().commitOrRollback(ctx, cache.cacheType, 793 encodedLen, 0, true, "") 794 } 795 cache.tlfCounts[tlfID]++ 796 cache.priorityBlockCounts[cache.homeDirs[tlfID]]++ 797 cache.priorityTlfMap[cache.homeDirs[tlfID]][tlfID]++ 798 cache.numBlocks++ 799 encodedLenUint := uint64(encodedLen) 800 cache.tlfSizes[tlfID] += encodedLenUint 801 cache.addCurrBytes(encodedLenUint) 802 } 803 tlfKey := cache.tlfKey(tlfID, blockKey) 804 hasKey, err = cache.tlfDb.Has(tlfKey, nil) 805 if err != nil { 806 cache.log.CWarningf(ctx, "Error reading from TLF cache database: %+v", 807 err) 808 } 809 if !hasKey { 810 err = cache.tlfDb.Put(tlfKey, []byte{}, nil) 811 if err != nil { 812 cache.log.CWarningf(ctx, 813 "Error writing to TLF cache database: %+v", err) 814 } 815 } 816 md, err := cache.getMetadataLocked(blockID, false) 817 if err != nil { 818 // Only set the relevant fields if we had trouble getting the metadata. 819 // Initially leave TriggeredPrefetch and FinishedPrefetch as false; 820 // rely on the later-called UpdateMetadata to fix it. 821 md.TlfID = tlfID 822 md.BlockSize = uint32(encodedLen) 823 err = nil 824 } 825 return cache.updateMetadataLocked(ctx, blockKey, md, ldbutils.Unmetered) 826 } 827 828 // GetMetadata implements the DiskBlockCache interface for 829 // DiskBlockCacheLocal. 830 func (cache *DiskBlockCacheLocal) GetMetadata(ctx context.Context, 831 blockID kbfsblock.ID) (DiskBlockCacheMetadata, error) { 832 cache.lock.RLock() 833 defer cache.lock.RUnlock() 834 err := cache.checkCacheLocked("Block(GetMetadata)") 835 if err != nil { 836 return DiskBlockCacheMetadata{}, err 837 } 838 return cache.getMetadataLocked(blockID, false) 839 } 840 841 // UpdateMetadata implements the DiskBlockCache interface for 842 // DiskBlockCacheLocal. 843 func (cache *DiskBlockCacheLocal) UpdateMetadata(ctx context.Context, 844 blockID kbfsblock.ID, prefetchStatus PrefetchStatus) (err error) { 845 cache.lock.Lock() 846 defer cache.lock.Unlock() 847 err = cache.checkCacheLocked("Block(UpdateMetadata)") 848 if err != nil { 849 return err 850 } 851 852 md, err := cache.getMetadataLocked(blockID, false) 853 if err != nil { 854 return data.NoSuchBlockError{ID: blockID} 855 } 856 if md.FinishedPrefetch { 857 // Don't update md that's already completed. 858 return nil 859 } 860 md.TriggeredPrefetch = false 861 md.FinishedPrefetch = false 862 switch prefetchStatus { 863 case TriggeredPrefetch: 864 md.TriggeredPrefetch = true 865 case FinishedPrefetch: 866 md.TriggeredPrefetch = true 867 md.FinishedPrefetch = true 868 } 869 return cache.updateMetadataLocked(ctx, blockID.Bytes(), md, ldbutils.Metered) 870 } 871 872 func (cache *DiskBlockCacheLocal) decCacheCountsLocked( 873 tlfID tlf.ID, numBlocks int, totalSize uint64) { 874 if numBlocks <= cache.tlfCounts[tlfID] { 875 cache.tlfCounts[tlfID] -= numBlocks 876 } 877 if numBlocks <= cache.priorityBlockCounts[cache.homeDirs[tlfID]] { 878 cache.priorityBlockCounts[cache.homeDirs[tlfID]] -= numBlocks 879 } 880 if numBlocks <= cache.priorityTlfMap[cache.homeDirs[tlfID]][tlfID] { 881 cache.priorityTlfMap[cache.homeDirs[tlfID]][tlfID] -= numBlocks 882 } 883 if numBlocks <= cache.numBlocks { 884 cache.numBlocks -= numBlocks 885 } 886 if totalSize <= cache.tlfSizes[tlfID] { 887 cache.tlfSizes[tlfID] -= totalSize 888 } 889 cache.subCurrBytes(totalSize) 890 } 891 892 // deleteLocked deletes a set of blocks from the disk block cache. 893 func (cache *DiskBlockCacheLocal) deleteLocked(ctx context.Context, 894 blockEntries []kbfsblock.ID) (numRemoved int, sizeRemoved int64, 895 err error) { 896 if len(blockEntries) == 0 { 897 return 0, 0, nil 898 } 899 defer func() { 900 if err == nil { 901 cache.deleteCountMeter.Mark(int64(numRemoved)) 902 cache.deleteSizeMeter.Mark(sizeRemoved) 903 } 904 }() 905 blockBatch := new(leveldb.Batch) 906 metadataBatch := new(leveldb.Batch) 907 tlfBatch := new(leveldb.Batch) 908 removalCounts := make(map[tlf.ID]int) 909 removalSizes := make(map[tlf.ID]uint64) 910 for _, entry := range blockEntries { 911 blockKey := entry.Bytes() 912 metadataBytes, err := cache.metaDb.Get(blockKey, nil) 913 if err != nil { 914 // If we can't retrieve the block, don't try to delete it, and 915 // don't account for its non-presence. 916 continue 917 } 918 metadata := DiskBlockCacheMetadata{} 919 err = cache.config.Codec().Decode(metadataBytes, &metadata) 920 if err != nil { 921 return 0, 0, err 922 } 923 blockBatch.Delete(blockKey) 924 metadataBatch.Delete(blockKey) 925 tlfDbKey := cache.tlfKey(metadata.TlfID, blockKey) 926 tlfBatch.Delete(tlfDbKey) 927 removalCounts[metadata.TlfID]++ 928 removalSizes[metadata.TlfID] += uint64(metadata.BlockSize) 929 sizeRemoved += int64(metadata.BlockSize) 930 numRemoved++ 931 } 932 // TODO: more gracefully handle non-atomic failures here. 933 if err := cache.metaDb.Write(metadataBatch, nil); err != nil { 934 return 0, 0, err 935 } 936 if err := cache.tlfDb.Write(tlfBatch, nil); err != nil { 937 return 0, 0, err 938 } 939 if err := cache.blockDb.Write(blockBatch, nil); err != nil { 940 return 0, 0, err 941 } 942 943 // Update the cache's totals. 944 for k, v := range removalCounts { 945 cache.decCacheCountsLocked(k, v, removalSizes[k]) 946 } 947 if cache.useLimiter() { 948 cache.config.DiskLimiter().release( 949 ctx, cache.cacheType, sizeRemoved, 0) 950 } 951 952 return numRemoved, sizeRemoved, nil 953 } 954 955 // compactDBs forces the disk cache to compact. It should be called 956 // after bulk deletion events, since level db doesn't run compactions 957 // after those (only after a certain number of lookups occur). 958 func (cache *DiskBlockCacheLocal) compactDBs(ctx context.Context) (err error) { 959 // We need to lock here to make sure the caches don't get shut 960 // down. TODO: make this less restrictive so this cache can still 961 // serve writes during compaction; however, I'm not yet sure if 962 // this is even allowed on the leveldb side or not. 963 cache.lock.RLock() 964 defer cache.lock.RUnlock() 965 err = cache.checkCacheLocked("compactDBs") 966 if err != nil { 967 return err 968 } 969 970 cache.log.CDebugf(ctx, "Compacting DBs for cacheType=%s", cache.cacheType) 971 defer func() { 972 cache.log.CDebugf( 973 ctx, "Done compacting DBs for cacheType=%s: %+v", cache.cacheType, 974 err) 975 }() 976 err = cache.blockDb.CompactRange(util.Range{}) 977 if err != nil { 978 return err 979 } 980 err = cache.tlfDb.CompactRange(util.Range{}) 981 if err != nil { 982 return err 983 } 984 return cache.metaDb.CompactRange(util.Range{}) 985 } 986 987 // compactLoops fires compaction, but only after five minutes of 988 // non-usage has passed. 989 func (cache *DiskBlockCacheLocal) compactLoop() { 990 ctx := context.Background() 991 var timer *time.Timer 992 var timerCh <-chan time.Time 993 for { 994 select { 995 case <-cache.compactCh: 996 // If we explicitly need to compact, start a new timer no 997 // matter what (since a compaction request implies a use 998 // of the cache). 999 if timer != nil { 1000 timer.Stop() 1001 } else { 1002 cache.log.CDebugf(ctx, "Starting initial compaction timer") 1003 } 1004 1005 timer = time.NewTimer(compactTimer) 1006 timerCh = timer.C 1007 case <-cache.useCh: 1008 // If we've just been used, interrupt any timer that's 1009 // already running, but don't start a new one if one isn't 1010 // already running. 1011 if timer != nil { 1012 timer.Stop() 1013 timer = time.NewTimer(compactTimer) 1014 timerCh = timer.C 1015 } 1016 case <-timerCh: 1017 err := cache.compactDBs(ctx) 1018 if err != nil { 1019 cache.log.CDebugf(ctx, "Error compacting DBs: %+v", err) 1020 } 1021 timerCh = nil 1022 timer = nil 1023 case <-cache.shutdownCh: 1024 close(cache.doneCh) 1025 return 1026 } 1027 } 1028 } 1029 1030 func (cache *DiskBlockCacheLocal) doCompact() { 1031 select { 1032 case cache.compactCh <- struct{}{}: 1033 default: 1034 } 1035 } 1036 1037 // Delete implements the DiskBlockCache interface for DiskBlockCacheLocal. 1038 func (cache *DiskBlockCacheLocal) Delete(ctx context.Context, 1039 blockIDs []kbfsblock.ID) (numRemoved int, sizeRemoved int64, err error) { 1040 defer func() { 1041 if err == nil && numRemoved > deleteCompactThreshold { 1042 cache.doCompact() 1043 } 1044 }() 1045 1046 cache.lock.Lock() 1047 defer cache.lock.Unlock() 1048 err = cache.checkCacheLocked("Block(Delete)") 1049 if err != nil { 1050 return 0, 0, err 1051 } 1052 1053 cache.log.CDebugf(ctx, "Cache Delete numBlocks=%d", len(blockIDs)) 1054 defer func() { 1055 cache.log.CDebugf(ctx, "Deleted numRequested=%d numRemoved=%d sizeRemoved=%d err=%+v", len(blockIDs), numRemoved, sizeRemoved, err) 1056 }() 1057 if cache.config.IsTestMode() { 1058 for _, bID := range blockIDs { 1059 cache.log.CDebugf(ctx, "Cache type=%d delete block ID %s", 1060 cache.cacheType, bID) 1061 } 1062 } 1063 return cache.deleteLocked(ctx, blockIDs) 1064 } 1065 1066 // getRandomBlockID gives us a pivot block ID for picking a random range of 1067 // blocks to consider deleting. We pick a point to start our range based on 1068 // the proportion of the TLF space taken up by numElements/totalElements. E.g. 1069 // if we need to consider 100 out of 400 blocks, and we assume that the block 1070 // IDs are uniformly distributed, then our random start point should be in the 1071 // [0,0.75) interval on the [0,1.0) block ID space. 1072 func (cache *DiskBlockCacheLocal) getRandomBlockID(numElements, 1073 totalElements int) (kbfsblock.ID, error) { 1074 if totalElements == 0 { 1075 return kbfsblock.ID{}, errors.New("") 1076 } 1077 // Return a 0 block ID pivot if we require more elements than the total 1078 // number available. 1079 if numElements >= totalElements { 1080 return kbfsblock.ID{}, nil 1081 } 1082 // Generate a random block ID to start the range. 1083 pivot := 1.0 - (float64(numElements) / float64(totalElements)) 1084 if cache.config.IsTestMode() { 1085 return kbfsblock.MakeRandomIDInRange(0, pivot, 1086 kbfsblock.UseMathRandForTest) 1087 } 1088 return kbfsblock.MakeRandomIDInRange(0, pivot, kbfsblock.UseRealRandomness) 1089 } 1090 1091 // evictSomeBlocks tries to evict `numBlocks` blocks from the cache. If 1092 // `blockIDs` doesn't have enough blocks, we evict them all and report how many 1093 // we evicted. 1094 func (cache *DiskBlockCacheLocal) evictSomeBlocks(ctx context.Context, 1095 numBlocks int, blockIDs blockIDsByTime) (numRemoved int, sizeRemoved int64, 1096 err error) { 1097 defer func() { 1098 cache.log.CDebugf(ctx, "Cache evictSomeBlocks numBlocksRequested=%d "+ 1099 "numBlocksEvicted=%d sizeBlocksEvicted=%d err=%+v", numBlocks, 1100 numRemoved, sizeRemoved, err) 1101 }() 1102 if len(blockIDs) <= numBlocks { 1103 numBlocks = len(blockIDs) 1104 } else { 1105 // Only sort if we need to grab a subset of blocks. 1106 sort.Sort(blockIDs) 1107 } 1108 1109 blocksToDelete := blockIDs.ToBlockIDSlice(numBlocks) 1110 return cache.deleteLocked(ctx, blocksToDelete) 1111 } 1112 1113 func (cache *DiskBlockCacheLocal) removeBrokenBlock( 1114 ctx context.Context, tlfID tlf.ID, blockID kbfsblock.ID) int64 { 1115 cache.log.CDebugf(ctx, "Removing broken block %s from the cache", blockID) 1116 blockKey := blockID.Bytes() 1117 entry, err := cache.blockDb.Get(blockKey, nil) 1118 if err != nil { 1119 cache.log.CDebugf(ctx, "Couldn't get %s: %+v", blockID, err) 1120 return 0 1121 } 1122 1123 err = cache.blockDb.Delete(blockKey, nil) 1124 if err != nil { 1125 cache.log.CDebugf(ctx, "Couldn't delete %s from block cache: %+v", 1126 blockID, err) 1127 return 0 1128 } 1129 1130 tlfKey := cache.tlfKey(tlfID, blockKey) 1131 err = cache.tlfDb.Delete(tlfKey, nil) 1132 if err != nil { 1133 cache.log.CWarningf(ctx, 1134 "Couldn't delete from TLF cache database: %+v", err) 1135 } 1136 1137 size := int64(len(entry)) 1138 // It's tough to know whether the block actually made it into the 1139 // block stats or not. If the block was added during this run of 1140 // KBFS, it will be in there; if it was loaded from disk, it 1141 // probably won't be in there, since the stats are loaded by 1142 // iterating over the metadata db. So it's very possible that 1143 // this will make the stats incorrect. ‾\_(ツ)_/‾. 1144 cache.decCacheCountsLocked(tlfID, 1, uint64(size)) 1145 if cache.useLimiter() { 1146 cache.config.DiskLimiter().release(ctx, cache.cacheType, size, 0) 1147 } 1148 1149 // Attempt to clean up corrupted metadata, if any. 1150 _ = cache.metaDb.Delete(blockKey, nil) 1151 return size 1152 } 1153 1154 // evictFromTLFLocked evicts a number of blocks from the cache for a given TLF. 1155 // We choose a pivot variable b randomly. Then begin an iterator into 1156 // cache.tlfDb.Range(tlfID + b, tlfID + MaxBlockID) and iterate from there to 1157 // get numBlocks * evictionConsiderationFactor block IDs. We sort the 1158 // resulting blocks by value (LRU time) and pick the minimum numBlocks. We then 1159 // call cache.Delete() on that list of block IDs. 1160 func (cache *DiskBlockCacheLocal) evictFromTLFLocked(ctx context.Context, 1161 tlfID tlf.ID, numBlocks int) (numRemoved int, sizeRemoved int64, err error) { 1162 tlfBytes := tlfID.Bytes() 1163 numElements := numBlocks * evictionConsiderationFactor 1164 blockID, err := cache.getRandomBlockID(numElements, cache.tlfCounts[tlfID]) 1165 if err != nil { 1166 return 0, 0, err 1167 } 1168 rng := &util.Range{ 1169 Start: append(tlfBytes, blockID.Bytes()...), 1170 Limit: append(tlfBytes, cache.maxBlockID...), 1171 } 1172 iter := cache.tlfDb.NewIterator(rng, nil) 1173 defer iter.Release() 1174 1175 blockIDs := make(blockIDsByTime, 0, numElements) 1176 var brokenIDs []kbfsblock.ID 1177 1178 for i := 0; i < numElements; i++ { 1179 if !iter.Next() { 1180 break 1181 } 1182 key := iter.Key() 1183 1184 blockIDBytes := key[len(tlfBytes):] 1185 blockID, err := kbfsblock.IDFromBytes(blockIDBytes) 1186 if err != nil { 1187 cache.log.CWarningf( 1188 ctx, "Error decoding block ID %x: %+v", blockIDBytes, err) 1189 brokenIDs = append(brokenIDs, blockID) 1190 continue 1191 } 1192 lru, err := cache.getLRULocked(blockID) 1193 if err != nil { 1194 cache.log.CWarningf( 1195 ctx, "Error decoding LRU time for block %s: %+v", blockID, err) 1196 brokenIDs = append(brokenIDs, blockID) 1197 continue 1198 } 1199 blockIDs = append(blockIDs, lruEntry{blockID, lru}) 1200 } 1201 1202 numRemoved, sizeRemoved, err = cache.evictSomeBlocks( 1203 ctx, numBlocks, blockIDs) 1204 if err != nil { 1205 return 0, 0, err 1206 } 1207 1208 for _, id := range brokenIDs { 1209 // Assume that a block that is in `tlfDB`, but for which the 1210 // metadata is missing or busted, 1211 size := cache.removeBrokenBlock(ctx, tlfID, id) 1212 if size > 0 { 1213 numRemoved++ 1214 sizeRemoved += size 1215 } 1216 } 1217 return numRemoved, sizeRemoved, nil 1218 } 1219 1220 // weightedByCount is used to shuffle TLF IDs, weighting by per-TLF block count. 1221 type weightedByCount struct { 1222 key float64 1223 value tlf.ID 1224 } 1225 1226 // shuffleTLFsAtPriorityWeighted shuffles the TLFs at a given priority, 1227 // weighting by per-TLF block count. 1228 func (cache *DiskBlockCacheLocal) shuffleTLFsAtPriorityWeighted( 1229 priority evictionPriority) []weightedByCount { 1230 weightedSlice := make([]weightedByCount, 0, 1231 len(cache.priorityTlfMap[priority])) 1232 idx := 0 1233 // Use an exponential distribution to ensure the weights are 1234 // correctly used. 1235 // See http://utopia.duth.gr/~pefraimi/research/data/2007EncOfAlg.pdf 1236 for tlfID, count := range cache.priorityTlfMap[priority] { 1237 if count == 0 { 1238 continue 1239 } 1240 weightedSlice = append(weightedSlice, weightedByCount{ 1241 key: math.Pow(rand.Float64(), 1.0/float64(count)), 1242 value: tlfID, 1243 }) 1244 idx++ 1245 } 1246 sort.Slice(weightedSlice, func(i, j int) bool { 1247 return weightedSlice[i].key > weightedSlice[j].key 1248 }) 1249 return weightedSlice 1250 } 1251 1252 // evictLocked evicts a number of blocks from the cache. We search the lowest 1253 // eviction priority level for blocks to evict first, then the next highest 1254 // priority and so on until enough blocks have been evicted. Within each 1255 // priority, we first shuffle the TLFs, weighting by how many blocks they 1256 // contain, and then we take the top TLFs from that shuffle and evict the 1257 // least recently used blocks from them. 1258 func (cache *DiskBlockCacheLocal) evictLocked(ctx context.Context, 1259 numBlocks int) (numRemoved int, sizeRemoved int64, err error) { 1260 numRemoved = 0 1261 sizeRemoved = 0 1262 defer func() { 1263 cache.evictCountMeter.Mark(int64(numRemoved)) 1264 cache.evictSizeMeter.Mark(sizeRemoved) 1265 }() 1266 for priorityToEvict := priorityNotHome; (priorityToEvict <= priorityPrivateHome) && (numRemoved < numBlocks); priorityToEvict++ { 1267 // Shuffle the TLFs of this priority, weighting by block count. 1268 shuffledSlice := cache.shuffleTLFsAtPriorityWeighted(priorityToEvict) 1269 // Select some TLFs to evict from. 1270 numElements := (numBlocks - numRemoved) * evictionConsiderationFactor 1271 1272 // blockIDs is a slice of blocks from which evictions will be selected. 1273 blockIDs := make(blockIDsByTime, 0, numElements) 1274 1275 // For each TLF until we get enough elements to select among, 1276 // add its blocks to the eviction slice. 1277 for _, tlfIDStruct := range shuffledSlice { 1278 tlfID := tlfIDStruct.value 1279 if cache.tlfCounts[tlfID] == 0 { 1280 cache.log.CDebugf(ctx, "No blocks to delete in TLF %s", tlfID) 1281 continue 1282 } 1283 tlfBytes := tlfID.Bytes() 1284 1285 blockID, err := cache.getRandomBlockID(numElements, 1286 cache.tlfCounts[tlfID]) 1287 if err != nil { 1288 return 0, 0, err 1289 } 1290 rng := &util.Range{ 1291 Start: append(tlfBytes, blockID.Bytes()...), 1292 Limit: append(tlfBytes, cache.maxBlockID...), 1293 } 1294 1295 // Extra func exists to make defers work. 1296 func() { 1297 iter := cache.tlfDb.NewIterator(rng, nil) 1298 defer iter.Release() 1299 1300 var brokenIDs []kbfsblock.ID 1301 for i := 0; i < numElements; i++ { 1302 if !iter.Next() { 1303 break 1304 } 1305 key := iter.Key() 1306 1307 blockIDBytes := key[len(tlfBytes):] 1308 blockID, err := kbfsblock.IDFromBytes(blockIDBytes) 1309 if err != nil { 1310 cache.log.CWarningf( 1311 ctx, "Error decoding block ID %x", blockIDBytes) 1312 brokenIDs = append(brokenIDs, blockID) 1313 continue 1314 } 1315 lru, err := cache.getLRULocked(blockID) 1316 if err != nil { 1317 cache.log.CWarningf( 1318 ctx, "Error decoding LRU time for block %s", 1319 blockID) 1320 brokenIDs = append(brokenIDs, blockID) 1321 continue 1322 } 1323 blockIDs = append(blockIDs, lruEntry{blockID, lru}) 1324 } 1325 1326 for _, id := range brokenIDs { 1327 // Assume that a block that is in `tlfDB`, but for which the 1328 // metadata is missing or busted, 1329 size := cache.removeBrokenBlock(ctx, tlfID, id) 1330 if size > 0 { 1331 numRemoved++ 1332 sizeRemoved += size 1333 } 1334 } 1335 }() 1336 if len(blockIDs) == numElements { 1337 break 1338 } 1339 } 1340 // Evict some of the selected blocks. 1341 currNumRemoved, currSizeRemoved, err := cache.evictSomeBlocks(ctx, 1342 numBlocks-numRemoved, blockIDs) 1343 if err != nil { 1344 return numRemoved, sizeRemoved, err 1345 } 1346 // Update the evicted count. 1347 numRemoved += currNumRemoved 1348 sizeRemoved += currSizeRemoved 1349 } 1350 1351 return numRemoved, sizeRemoved, nil 1352 } 1353 1354 func (cache *DiskBlockCacheLocal) deleteNextBatchFromClearedTlf( 1355 ctx context.Context, tlfID tlf.ID) (numLeft int, err error) { 1356 cache.lock.Lock() 1357 defer cache.lock.Unlock() 1358 err = cache.checkCacheLocked("Block(deleteNextBatchFromClearedTlf)") 1359 if err != nil { 1360 return 0, err 1361 } 1362 1363 select { 1364 case <-ctx.Done(): 1365 return 0, ctx.Err() 1366 default: 1367 } 1368 1369 _, _, err = cache.evictFromTLFLocked( 1370 ctx, tlfID, cache.numBlocksToEvictOnClear) 1371 if err != nil { 1372 return 0, err 1373 } 1374 return cache.tlfCounts[tlfID], nil 1375 } 1376 1377 // ClearAllTlfBlocks implements the DiskBlockCache interface for 1378 // DiskBlockCacheLocal. 1379 func (cache *DiskBlockCacheLocal) ClearAllTlfBlocks( 1380 ctx context.Context, tlfID tlf.ID) (err error) { 1381 defer func() { 1382 cache.log.CDebugf(ctx, 1383 "Finished clearing blocks from %s: %+v", tlfID, err) 1384 if err == nil { 1385 cache.doCompact() 1386 } 1387 }() 1388 1389 // Delete the blocks in batches, so we don't keep the lock for too 1390 // long. 1391 for { 1392 cache.log.CDebugf(ctx, "Deleting a batch of blocks from %s", tlfID) 1393 numLeft, err := cache.deleteNextBatchFromClearedTlf(ctx, tlfID) 1394 if err != nil { 1395 return err 1396 } 1397 if numLeft == 0 { 1398 cache.log.CDebugf(ctx, "Deleted all blocks from %s", tlfID) 1399 return nil 1400 } 1401 cache.log.CDebugf( 1402 ctx, "%d blocks left to delete from %s", numLeft, tlfID) 1403 1404 c := time.After(cache.clearTickerDuration) 1405 select { 1406 case <-c: 1407 case <-ctx.Done(): 1408 return ctx.Err() 1409 } 1410 } 1411 } 1412 1413 // GetLastUnrefRev implements the DiskBlockCache interface for 1414 // DiskBlockCacheLocal. 1415 func (cache *DiskBlockCacheLocal) GetLastUnrefRev( 1416 ctx context.Context, tlfID tlf.ID) (kbfsmd.Revision, error) { 1417 cache.lock.RLock() 1418 defer cache.lock.RUnlock() 1419 err := cache.checkCacheLocked("Block(GetLastUnrefRev)") 1420 if err != nil { 1421 return kbfsmd.RevisionUninitialized, err 1422 } 1423 1424 rev, ok := cache.tlfLastUnrefs[tlfID] 1425 if !ok { 1426 // No known unref'd revision. 1427 return kbfsmd.RevisionUninitialized, nil 1428 } 1429 return rev, nil 1430 } 1431 1432 // PutLastUnrefRev implements the DiskBlockCache interface for 1433 // DiskBlockCacheLocal. 1434 func (cache *DiskBlockCacheLocal) PutLastUnrefRev( 1435 ctx context.Context, tlfID tlf.ID, rev kbfsmd.Revision) error { 1436 cache.lock.Lock() 1437 defer cache.lock.Unlock() 1438 err := cache.checkCacheLocked("Block(PutLastUnrefRev)") 1439 if err != nil { 1440 return err 1441 } 1442 1443 if currRev, ok := cache.tlfLastUnrefs[tlfID]; ok { 1444 if rev <= currRev { 1445 // A later revision has already been unref'd, so ignore this. 1446 return nil 1447 } 1448 } 1449 1450 buf, err := cache.encodeLastUnref(rev) 1451 if err != nil { 1452 return err 1453 } 1454 err = cache.lastUnrefDb.Put(tlfID.Bytes(), buf, nil) 1455 if err != nil { 1456 return err 1457 } 1458 cache.tlfLastUnrefs[tlfID] = rev 1459 return nil 1460 } 1461 1462 // Status implements the DiskBlockCache interface for DiskBlockCacheLocal. 1463 func (cache *DiskBlockCacheLocal) Status( 1464 ctx context.Context) map[string]DiskBlockCacheStatus { 1465 var name string 1466 var maxLimit uint64 1467 limiterStatus := cache.config.DiskLimiter().getStatus( 1468 ctx, keybase1.UserOrTeamID("")).(backpressureDiskLimiterStatus) 1469 switch cache.cacheType { 1470 case syncCacheLimitTrackerType: 1471 name = syncCacheName 1472 maxLimit = uint64(limiterStatus.SyncCacheByteStatus.Max) 1473 case workingSetCacheLimitTrackerType: 1474 name = workingSetCacheName 1475 maxLimit = uint64(limiterStatus.DiskCacheByteStatus.Max) 1476 case crDirtyBlockCacheLimitTrackerType: 1477 name = crDirtyBlockCacheName 1478 } 1479 select { 1480 case <-cache.startedCh: 1481 case <-cache.startErrCh: 1482 return map[string]DiskBlockCacheStatus{name: {StartState: DiskBlockCacheStartStateFailed}} 1483 default: 1484 return map[string]DiskBlockCacheStatus{name: {StartState: DiskBlockCacheStartStateStarting}} 1485 } 1486 availableBytes, totalBytes := uint64(math.MaxInt64), uint64(math.MaxInt64) 1487 if cache.dirPath != "" { 1488 var err error 1489 availableBytes, totalBytes, _, _, err = getDiskLimits(cache.dirPath) 1490 if err != nil { 1491 cache.log.CDebugf(ctx, "Couldn't get disk stats: %+v", err) 1492 } 1493 } 1494 1495 cache.lock.RLock() 1496 defer cache.lock.RUnlock() 1497 1498 var blockStats, metaStats, tlfStats, lastUnrefStats []string 1499 var memCompActive, tableCompActive bool 1500 var metaMemCompActive, metaTableCompActive bool 1501 if err := cache.checkCacheLocked("Block(Status)"); err == nil { 1502 blockStats, err = cache.blockDb.StatStrings() 1503 if err != nil { 1504 cache.log.CDebugf(ctx, "Couldn't get block db stats: %+v", err) 1505 } 1506 metaStats, err = cache.metaDb.StatStrings() 1507 if err != nil { 1508 cache.log.CDebugf(ctx, "Couldn't get meta db stats: %+v", err) 1509 } 1510 tlfStats, err = cache.tlfDb.StatStrings() 1511 if err != nil { 1512 cache.log.CDebugf(ctx, "Couldn't get TLF db stats: %+v", err) 1513 } 1514 lastUnrefStats, err = cache.lastUnrefDb.StatStrings() 1515 if err != nil { 1516 cache.log.CDebugf(ctx, "Couldn't get last unref db stats: %+v", err) 1517 } 1518 var dbStats leveldb.DBStats 1519 err = cache.blockDb.Stats(&dbStats) 1520 if err != nil { 1521 cache.log.CDebugf( 1522 ctx, "Couldn't get block db compaction stats: %+v", err) 1523 } 1524 memCompActive, tableCompActive = 1525 dbStats.MemCompactionActive, dbStats.TableCompactionActive 1526 err = cache.metaDb.Stats(&dbStats) 1527 if err != nil { 1528 cache.log.CDebugf( 1529 ctx, "Couldn't get meta db compaction stats: %+v", err) 1530 } 1531 metaMemCompActive, metaTableCompActive = 1532 dbStats.MemCompactionActive, dbStats.TableCompactionActive 1533 } 1534 1535 // The disk cache status doesn't depend on the chargedTo ID, and 1536 // we don't have easy access to the UID here, so pass in a dummy. 1537 return map[string]DiskBlockCacheStatus{ 1538 name: { 1539 StartState: DiskBlockCacheStartStateStarted, 1540 NumBlocks: uint64(cache.numBlocks), 1541 BlockBytes: cache.getCurrBytes(), 1542 CurrByteLimit: maxLimit, 1543 LastUnrefCount: uint64(len(cache.tlfLastUnrefs)), 1544 Hits: ldbutils.RateMeterToStatus(cache.hitMeter), 1545 Misses: ldbutils.RateMeterToStatus(cache.missMeter), 1546 Puts: ldbutils.RateMeterToStatus(cache.putMeter), 1547 MetadataUpdates: ldbutils.RateMeterToStatus(cache.updateMeter), 1548 NumEvicted: ldbutils.RateMeterToStatus(cache.evictCountMeter), 1549 SizeEvicted: ldbutils.RateMeterToStatus(cache.evictSizeMeter), 1550 NumDeleted: ldbutils.RateMeterToStatus(cache.deleteCountMeter), 1551 SizeDeleted: ldbutils.RateMeterToStatus(cache.deleteSizeMeter), 1552 LocalDiskBytesAvailable: availableBytes, 1553 LocalDiskBytesTotal: totalBytes, 1554 BlockDBStats: blockStats, 1555 MetaDBStats: metaStats, 1556 MemCompActive: memCompActive, 1557 TableCompActive: tableCompActive, 1558 MetaMemCompActive: metaMemCompActive, 1559 MetaTableCompActive: metaTableCompActive, 1560 TLFDBStats: tlfStats, 1561 LastUnrefStats: lastUnrefStats, 1562 }, 1563 } 1564 } 1565 1566 // DoesCacheHaveSpace returns true if we have more than 1% of space 1567 // left in the cache. 1568 func (cache *DiskBlockCacheLocal) DoesCacheHaveSpace( 1569 ctx context.Context) (hasSpace bool, howMuch int64, err error) { 1570 limiterStatus := cache.config.DiskLimiter().getStatus( 1571 ctx, keybase1.UserOrTeamID("")).(backpressureDiskLimiterStatus) 1572 switch cache.cacheType { 1573 case syncCacheLimitTrackerType: 1574 // The tracker doesn't track sync cache usage because we never 1575 // want to throttle it, so rely on our local byte usage count 1576 // instead of the fraction returned by the tracker. 1577 limit := float64(limiterStatus.SyncCacheByteStatus.Max) 1578 return float64(cache.getCurrBytes())/limit <= .99, 1579 limiterStatus.DiskCacheByteStatus.Free, nil 1580 case workingSetCacheLimitTrackerType: 1581 return limiterStatus.DiskCacheByteStatus.UsedFrac <= .99, 1582 limiterStatus.DiskCacheByteStatus.Free, nil 1583 case crDirtyBlockCacheLimitTrackerType: 1584 return true, 0, nil 1585 default: 1586 panic(fmt.Sprintf("Unknown cache type: %d", cache.cacheType)) 1587 } 1588 } 1589 1590 // Mark updates the metadata of the given block with the tag. 1591 func (cache *DiskBlockCacheLocal) Mark( 1592 ctx context.Context, blockID kbfsblock.ID, tag string) error { 1593 cache.lock.Lock() 1594 defer cache.lock.Unlock() 1595 err := cache.checkCacheLocked("Block(UpdateMetadata)") 1596 if err != nil { 1597 return err 1598 } 1599 1600 md, err := cache.getMetadataLocked(blockID, false) 1601 if err != nil { 1602 return data.NoSuchBlockError{ID: blockID} 1603 } 1604 md.Tag = tag 1605 return cache.updateMetadataLocked(ctx, blockID.Bytes(), md, false) 1606 } 1607 1608 func (cache *DiskBlockCacheLocal) deleteNextUnmarkedBatchFromTlf( 1609 ctx context.Context, tlfID tlf.ID, tag string, startingKey []byte) ( 1610 nextKey []byte, err error) { 1611 cache.lock.Lock() 1612 defer cache.lock.Unlock() 1613 err = cache.checkCacheLocked("Block(deleteNextUnmarkedBatchFromTlf)") 1614 if err != nil { 1615 return nil, err 1616 } 1617 1618 select { 1619 case <-ctx.Done(): 1620 return nil, ctx.Err() 1621 default: 1622 } 1623 1624 tlfBytes := tlfID.Bytes() 1625 rng := &util.Range{ 1626 Start: startingKey, 1627 Limit: append(tlfBytes, cache.maxBlockID...), 1628 } 1629 iter := cache.tlfDb.NewIterator(rng, nil) 1630 defer iter.Release() 1631 1632 blockIDs := make([]kbfsblock.ID, 0, cache.numUnmarkedBlocksToCheck) 1633 for i := 0; i < cache.numUnmarkedBlocksToCheck; i++ { 1634 if !iter.Next() { 1635 break 1636 } 1637 key := iter.Key() 1638 1639 blockIDBytes := key[len(tlfBytes):] 1640 blockID, err := kbfsblock.IDFromBytes(blockIDBytes) 1641 if err != nil { 1642 cache.log.CWarningf(ctx, "Error decoding block ID %x", blockIDBytes) 1643 continue 1644 } 1645 md, err := cache.getMetadataLocked(blockID, false) 1646 if err != nil { 1647 cache.log.CWarningf( 1648 ctx, "No metadata for %s while checking mark", blockID) 1649 continue 1650 } 1651 if md.Tag != tag { 1652 blockIDs = append(blockIDs, blockID) 1653 } 1654 } 1655 1656 if iter.Next() { 1657 nextKey = iter.Key() 1658 } 1659 1660 if len(blockIDs) > 0 { 1661 cache.log.CDebugf(ctx, "Deleting %d unmarked blocks (tag=%s) from %s", 1662 len(blockIDs), tag, tlfID) 1663 _, _, err = cache.deleteLocked(ctx, blockIDs) 1664 if err != nil { 1665 return nil, err 1666 } 1667 } 1668 return nextKey, nil 1669 } 1670 1671 // DeleteUnmarked deletes all the blocks without the given tag. 1672 func (cache *DiskBlockCacheLocal) DeleteUnmarked( 1673 ctx context.Context, tlfID tlf.ID, tag string) (err error) { 1674 defer func() { 1675 cache.log.CDebugf(ctx, 1676 "Finished deleting unmarked blocks (tag=%s) from %s: %+v", 1677 tag, tlfID, err) 1678 if err == nil { 1679 cache.doCompact() 1680 } 1681 }() 1682 1683 // Delete the blocks in batches, so we don't keep the lock for too 1684 // long. 1685 startingKey := cache.tlfKey(tlfID, nil) 1686 for { 1687 cache.log.CDebugf( 1688 ctx, "Deleting a batch of unmarked blocks (tag=%s) from %s", 1689 tag, tlfID) 1690 startingKey, err = cache.deleteNextUnmarkedBatchFromTlf( 1691 ctx, tlfID, tag, startingKey) 1692 if err != nil { 1693 return err 1694 } 1695 if startingKey == nil { 1696 return nil 1697 } 1698 1699 c := time.After(cache.clearTickerDuration) 1700 select { 1701 case <-c: 1702 case <-ctx.Done(): 1703 return ctx.Err() 1704 } 1705 } 1706 } 1707 1708 // AddHomeTLF implements this DiskBlockCache interace for DiskBlockCacheLocal. 1709 func (cache *DiskBlockCacheLocal) AddHomeTLF(ctx context.Context, tlfID tlf.ID) error { 1710 cache.lock.Lock() 1711 defer cache.lock.Unlock() 1712 cache.priorityBlockCounts[cache.homeDirs[tlfID]] -= cache.tlfCounts[tlfID] 1713 cache.priorityTlfMap[cache.homeDirs[tlfID]][tlfID] -= cache.tlfCounts[tlfID] 1714 1715 switch tlfID.Type() { 1716 case tlf.Private: 1717 cache.homeDirs[tlfID] = priorityPrivateHome 1718 case tlf.Public: 1719 cache.homeDirs[tlfID] = priorityPublicHome 1720 default: 1721 return errTeamOrUnknownTLFAddedAsHome 1722 } 1723 cache.priorityBlockCounts[cache.homeDirs[tlfID]] += cache.tlfCounts[tlfID] 1724 cache.priorityTlfMap[cache.homeDirs[tlfID]][tlfID] += cache.tlfCounts[tlfID] 1725 1726 return nil 1727 } 1728 1729 // ClearHomeTLFs implements this DiskBlockCache interace for 1730 // DiskBlockCacheLocal. 1731 func (cache *DiskBlockCacheLocal) ClearHomeTLFs(ctx context.Context) error { 1732 cache.lock.Lock() 1733 defer cache.lock.Unlock() 1734 for tlfID, priority := range cache.homeDirs { 1735 cache.priorityBlockCounts[priority] -= cache.tlfCounts[tlfID] 1736 cache.priorityTlfMap[priority][tlfID] -= cache.tlfCounts[tlfID] 1737 cache.priorityBlockCounts[priorityNotHome] += cache.tlfCounts[tlfID] 1738 cache.priorityTlfMap[priorityNotHome][tlfID] += cache.tlfCounts[tlfID] 1739 } 1740 cache.homeDirs = make(map[tlf.ID]evictionPriority) 1741 return nil 1742 } 1743 1744 // GetTlfSize returns the number of bytes stored for the given TLF in 1745 // the cache. 1746 func (cache *DiskBlockCacheLocal) GetTlfSize( 1747 _ context.Context, tlfID tlf.ID) (uint64, error) { 1748 cache.lock.RLock() 1749 defer cache.lock.RUnlock() 1750 return cache.tlfSizes[tlfID], nil 1751 } 1752 1753 // GetTlfIDs returns the IDs of all the TLFs with blocks stored in 1754 // the cache. 1755 func (cache *DiskBlockCacheLocal) GetTlfIDs( 1756 _ context.Context) (tlfIDs []tlf.ID, err error) { 1757 cache.lock.RLock() 1758 defer cache.lock.RUnlock() 1759 tlfIDs = make([]tlf.ID, 0, len(cache.tlfSizes)) 1760 for id := range cache.tlfSizes { 1761 tlfIDs = append(tlfIDs, id) 1762 } 1763 return tlfIDs, nil 1764 } 1765 1766 // Shutdown implements the DiskBlockCache interface for DiskBlockCacheLocal. 1767 func (cache *DiskBlockCacheLocal) Shutdown(ctx context.Context) <-chan struct{} { 1768 // Wait for the cache to either finish starting or error. 1769 select { 1770 case <-cache.startedCh: 1771 case <-cache.startErrCh: 1772 close(cache.doneCh) 1773 return cache.doneCh 1774 } 1775 cache.lock.Lock() 1776 defer cache.lock.Unlock() 1777 // shutdownCh has to be checked under lock, otherwise we can race. 1778 select { 1779 case <-cache.shutdownCh: 1780 cache.log.CWarningf(ctx, "Shutdown called more than once") 1781 return cache.doneCh 1782 default: 1783 } 1784 close(cache.shutdownCh) 1785 if cache.blockDb == nil { 1786 return cache.doneCh 1787 } 1788 cache.closer() 1789 cache.blockDb = nil 1790 cache.metaDb = nil 1791 cache.tlfDb = nil 1792 if cache.useLimiter() { 1793 cache.config.DiskLimiter().onSimpleByteTrackerDisable(ctx, 1794 cache.cacheType, int64(cache.getCurrBytes())) 1795 } 1796 cache.hitMeter.Shutdown() 1797 cache.missMeter.Shutdown() 1798 cache.putMeter.Shutdown() 1799 cache.updateMeter.Shutdown() 1800 cache.evictCountMeter.Shutdown() 1801 cache.evictSizeMeter.Shutdown() 1802 cache.deleteCountMeter.Shutdown() 1803 cache.deleteSizeMeter.Shutdown() 1804 return cache.doneCh 1805 }