github.com/m3db/m3@v1.5.0/src/dbnode/persist/fs/seek_manager_test.go (about) 1 // Copyright (c) 2016 Uber Technologies, Inc. 2 // 3 // Permission is hereby granted, free of charge, to any person obtaining a copy 4 // of this software and associated documentation files (the "Software"), to deal 5 // in the Software without restriction, including without limitation the rights 6 // to use, copy, modify, merge, publish, distribute, sublicense, and/or sell 7 // copies of the Software, and to permit persons to whom the Software is 8 // furnished to do so, subject to the following conditions: 9 // 10 // The above copyright notice and this permission notice shall be included in 11 // all copies or substantial portions of the Software. 12 // 13 // THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR 14 // IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, 15 // FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE 16 // AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER 17 // LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, 18 // OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN 19 // THE SOFTWARE. 20 21 package fs 22 23 import ( 24 "sync" 25 "testing" 26 "time" 27 28 "github.com/m3db/m3/src/cluster/shard" 29 "github.com/m3db/m3/src/dbnode/retention" 30 "github.com/m3db/m3/src/dbnode/sharding" 31 "github.com/m3db/m3/src/dbnode/storage/block" 32 "github.com/m3db/m3/src/x/ident" 33 xtest "github.com/m3db/m3/src/x/test" 34 xtime "github.com/m3db/m3/src/x/time" 35 36 "github.com/fortytw2/leaktest" 37 "github.com/golang/mock/gomock" 38 "github.com/stretchr/testify/require" 39 ) 40 41 const ( 42 defaultTestingFetchConcurrency = 2 43 ) 44 45 var defaultTestBlockRetrieverOptions = NewBlockRetrieverOptions(). 46 SetBlockLeaseManager(&block.NoopLeaseManager{}). 47 // Test with caching enabled. 48 SetCacheBlocksOnRetrieve(true). 49 // Default value is determined by available CPUs, but for testing 50 // we want to have this been consistent across hardware. 51 SetFetchConcurrency(defaultTestingFetchConcurrency) 52 53 func TestSeekerManagerCacheShardIndices(t *testing.T) { 54 defer leaktest.CheckTimeout(t, 1*time.Minute)() 55 56 shards := []uint32{2, 5, 9, 478, 1023} 57 metadata := testNs1Metadata(t) 58 shardSet, err := sharding.NewShardSet( 59 sharding.NewShards(shards, shard.Available), 60 sharding.DefaultHashFn(1), 61 ) 62 require.NoError(t, err) 63 m := NewSeekerManager(nil, testDefaultOpts, defaultTestBlockRetrieverOptions).(*seekerManager) 64 require.NoError(t, m.Open(metadata, shardSet)) 65 byTimes := make(map[uint32]*seekersByTime) 66 var mu sync.Mutex 67 m.openAnyUnopenSeekersFn = func(byTime *seekersByTime) error { 68 mu.Lock() 69 byTimes[byTime.shard] = byTime 70 mu.Unlock() 71 return nil 72 } 73 74 require.NoError(t, m.CacheShardIndices(shards)) 75 // Assert captured byTime objects match expectations 76 require.Equal(t, len(shards), len(byTimes)) 77 for _, shard := range shards { 78 mu.Lock() 79 byTimes[shard].shard = shard 80 mu.Unlock() 81 } 82 83 // Assert seeksByShardIdx match expectations 84 shardSetMap := make(map[uint32]struct{}, len(shards)) 85 for _, shard := range shards { 86 shardSetMap[shard] = struct{}{} 87 } 88 89 for shard, byTime := range m.seekersByShardIdx { 90 _, exists := shardSetMap[uint32(shard)] 91 if !exists { 92 require.False(t, byTime.accessed) 93 } else { 94 require.True(t, byTime.accessed) 95 require.Equal(t, int(shard), int(byTime.shard)) 96 } 97 } 98 99 require.NoError(t, m.Close()) 100 } 101 102 func TestSeekerManagerUpdateOpenLease(t *testing.T) { 103 defer leaktest.CheckTimeout(t, 1*time.Minute)() 104 105 var ( 106 ctrl = xtest.NewController(t) 107 shards = []uint32{2, 5, 9, 478, 1023} 108 m = NewSeekerManager(nil, testDefaultOpts, defaultTestBlockRetrieverOptions).(*seekerManager) 109 ) 110 defer ctrl.Finish() 111 112 var ( 113 mockSeekerStatsLock sync.Mutex 114 numMockSeekerCloses int 115 ) 116 m.newOpenSeekerFn = func( 117 shard uint32, 118 blockStart xtime.UnixNano, 119 volume int, 120 ) (DataFileSetSeeker, error) { 121 mock := NewMockDataFileSetSeeker(ctrl) 122 // ConcurrentClone() will be called fetchConcurrency-1 times because the original can be used 123 // as one of the clones. 124 for i := 0; i < defaultTestingFetchConcurrency-1; i++ { 125 mock.EXPECT().ConcurrentClone().Return(mock, nil) 126 } 127 for i := 0; i < defaultTestingFetchConcurrency; i++ { 128 mock.EXPECT().Close().DoAndReturn(func() error { 129 mockSeekerStatsLock.Lock() 130 numMockSeekerCloses++ 131 mockSeekerStatsLock.Unlock() 132 return nil 133 }) 134 mock.EXPECT().ConcurrentIDBloomFilter().Return(nil).AnyTimes() 135 } 136 return mock, nil 137 } 138 m.sleepFn = func(_ time.Duration) { 139 time.Sleep(time.Millisecond) 140 } 141 142 metadata := testNs1Metadata(t) 143 shardSet, err := sharding.NewShardSet( 144 sharding.NewShards(shards, shard.Available), 145 sharding.DefaultHashFn(1), 146 ) 147 require.NoError(t, err) 148 // Pick a start time that's within retention so the background loop doesn't close 149 // the seeker. 150 blockStart := xtime.Now().Truncate(metadata.Options().RetentionOptions().BlockSize()) 151 require.NoError(t, m.Open(metadata, shardSet)) 152 for _, shard := range shards { 153 seeker, err := m.Borrow(shard, blockStart) 154 require.NoError(t, err) 155 byTime, ok := m.seekersByTime(shard) 156 require.True(t, ok) 157 byTime.RLock() 158 seekers := byTime.seekers[blockStart] 159 require.Equal(t, defaultTestingFetchConcurrency, len(seekers.active.seekers)) 160 require.Equal(t, 0, seekers.active.volume) 161 byTime.RUnlock() 162 require.NoError(t, m.Return(shard, blockStart, seeker)) 163 } 164 165 // Ensure that UpdateOpenLease() updates the volumes. 166 for _, shard := range shards { 167 updateResult, err := m.UpdateOpenLease(block.LeaseDescriptor{ 168 Namespace: metadata.ID(), 169 Shard: shard, 170 BlockStart: blockStart, 171 }, block.LeaseState{Volume: 1}) 172 require.NoError(t, err) 173 require.Equal(t, block.UpdateOpenLease, updateResult) 174 175 byTime, ok := m.seekersByTime(shard) 176 require.True(t, ok) 177 byTime.RLock() 178 seekers := byTime.seekers[blockStart] 179 require.Equal(t, defaultTestingFetchConcurrency, len(seekers.active.seekers)) 180 require.Equal(t, 1, seekers.active.volume) 181 byTime.RUnlock() 182 } 183 // Ensure that the old seekers actually get closed. 184 mockSeekerStatsLock.Lock() 185 require.Equal(t, len(shards)*defaultTestingFetchConcurrency, numMockSeekerCloses) 186 mockSeekerStatsLock.Unlock() 187 188 // Ensure that UpdateOpenLease() ignores updates for the wrong namespace. 189 for _, shard := range shards { 190 updateResult, err := m.UpdateOpenLease(block.LeaseDescriptor{ 191 Namespace: ident.StringID("some-other-ns"), 192 Shard: shard, 193 BlockStart: blockStart, 194 }, block.LeaseState{Volume: 2}) 195 require.NoError(t, err) 196 require.Equal(t, block.NoOpenLease, updateResult) 197 198 byTime, ok := m.seekersByTime(shard) 199 require.True(t, ok) 200 byTime.RLock() 201 seekers := byTime.seekers[blockStart] 202 require.Equal(t, defaultTestingFetchConcurrency, len(seekers.active.seekers)) 203 // Should not have increased to 2. 204 require.Equal(t, 1, seekers.active.volume) 205 byTime.RUnlock() 206 } 207 208 // Ensure that UpdateOpenLease() returns an error for out-of-order updates. 209 for _, shard := range shards { 210 _, err := m.UpdateOpenLease(block.LeaseDescriptor{ 211 Namespace: metadata.ID(), 212 Shard: shard, 213 BlockStart: blockStart, 214 }, block.LeaseState{Volume: 0}) 215 require.Equal(t, errOutOfOrderUpdateOpenLease, err) 216 } 217 218 require.NoError(t, m.Close()) 219 } 220 221 func TestSeekerManagerUpdateOpenLeaseConcurrentNotAllowed(t *testing.T) { 222 defer leaktest.CheckTimeout(t, 1*time.Minute)() 223 224 var ( 225 ctrl = xtest.NewController(t) 226 shards = []uint32{1, 2} 227 m = NewSeekerManager(nil, testDefaultOpts, defaultTestBlockRetrieverOptions).(*seekerManager) 228 metadata = testNs1Metadata(t) 229 // Pick a start time that's within retention so the background loop doesn't close the seeker. 230 blockStart = xtime.Now().Truncate(metadata.Options().RetentionOptions().BlockSize()) 231 ) 232 defer ctrl.Finish() 233 234 descriptor1 := block.LeaseDescriptor{ 235 Namespace: metadata.ID(), 236 Shard: 1, 237 BlockStart: blockStart, 238 } 239 240 m.newOpenSeekerFn = func( 241 shard uint32, 242 blockStart xtime.UnixNano, 243 volume int, 244 ) (DataFileSetSeeker, error) { 245 if volume == 1 { 246 var wg sync.WaitGroup 247 wg.Add(1) 248 go func() { 249 defer wg.Done() 250 // Call UpdateOpenLease while within another UpdateOpenLease call. 251 _, err := m.UpdateOpenLease(descriptor1, block.LeaseState{Volume: 2}) 252 if shard == 1 { 253 // Concurrent call is made with the same shard id (and other values). 254 require.Equal(t, errConcurrentUpdateOpenLeaseNotAllowed, err) 255 } else { 256 // Concurrent call is made with a different shard id (2) and so it should pass. 257 require.NoError(t, err) 258 } 259 }() 260 wg.Wait() 261 } 262 mock := NewMockDataFileSetSeeker(ctrl) 263 mock.EXPECT().ConcurrentClone().Return(mock, nil).AnyTimes() 264 mock.EXPECT().Close().AnyTimes() 265 mock.EXPECT().ConcurrentIDBloomFilter().Return(nil).AnyTimes() 266 return mock, nil 267 } 268 m.sleepFn = func(_ time.Duration) { 269 time.Sleep(time.Millisecond) 270 } 271 272 shardSet, err := sharding.NewShardSet( 273 sharding.NewShards(shards, shard.Available), 274 sharding.DefaultHashFn(1), 275 ) 276 require.NoError(t, err) 277 require.NoError(t, m.Open(metadata, shardSet)) 278 279 for _, shardID := range shards { 280 seeker, err := m.Borrow(shardID, blockStart) 281 require.NoError(t, err) 282 require.NoError(t, m.Return(shardID, blockStart, seeker)) 283 } 284 285 updateResult, err := m.UpdateOpenLease(descriptor1, block.LeaseState{Volume: 1}) 286 require.NoError(t, err) 287 require.Equal(t, block.UpdateOpenLease, updateResult) 288 289 descriptor2 := descriptor1 290 descriptor2.Shard = 2 291 updateResult, err = m.UpdateOpenLease(descriptor2, block.LeaseState{Volume: 1}) 292 require.NoError(t, err) 293 require.Equal(t, block.UpdateOpenLease, updateResult) 294 295 require.NoError(t, m.Close()) 296 } 297 298 // TestSeekerManagerBorrowOpenSeekersLazy tests that the Borrow() method will 299 // open seekers lazily if they're not already open. 300 func TestSeekerManagerBorrowOpenSeekersLazy(t *testing.T) { 301 defer leaktest.CheckTimeout(t, 1*time.Minute)() 302 303 ctrl := xtest.NewController(t) 304 305 shards := []uint32{2, 5, 9, 478, 1023} 306 m := NewSeekerManager(nil, testDefaultOpts, defaultTestBlockRetrieverOptions).(*seekerManager) 307 m.newOpenSeekerFn = func( 308 shard uint32, 309 blockStart xtime.UnixNano, 310 volume int, 311 ) (DataFileSetSeeker, error) { 312 mock := NewMockDataFileSetSeeker(ctrl) 313 mock.EXPECT().Open(gomock.Any(), gomock.Any(), gomock.Any(), gomock.Any(), gomock.Any()).Return(nil) 314 mock.EXPECT().ConcurrentClone().Return(mock, nil) 315 for i := 0; i < defaultTestingFetchConcurrency; i++ { 316 mock.EXPECT().Close().Return(nil) 317 mock.EXPECT().ConcurrentIDBloomFilter().Return(nil) 318 } 319 return mock, nil 320 } 321 m.sleepFn = func(_ time.Duration) { 322 time.Sleep(time.Millisecond) 323 } 324 325 metadata := testNs1Metadata(t) 326 shardSet, err := sharding.NewShardSet( 327 sharding.NewShards(shards, shard.Available), 328 sharding.DefaultHashFn(1), 329 ) 330 require.NoError(t, err) 331 require.NoError(t, m.Open(metadata, shardSet)) 332 for _, shard := range shards { 333 seeker, err := m.Borrow(shard, 0) 334 require.NoError(t, err) 335 byTime, ok := m.seekersByTime(shard) 336 require.True(t, ok) 337 byTime.RLock() 338 seekers := byTime.seekers[0] 339 require.Equal(t, defaultTestingFetchConcurrency, len(seekers.active.seekers)) 340 byTime.RUnlock() 341 require.NoError(t, m.Return(shard, 0, seeker)) 342 } 343 344 require.NoError(t, m.Close()) 345 } 346 347 // TestSeekerManagerOpenCloseLoop tests the openCloseLoop of the SeekerManager 348 // by making sure that it makes the right decisions with regards to cleaning 349 // up resources based on their state. 350 func TestSeekerManagerOpenCloseLoop(t *testing.T) { 351 defer leaktest.CheckTimeout(t, 1*time.Minute)() 352 353 ctrl := xtest.NewController(t) 354 m := NewSeekerManager(nil, testDefaultOpts, defaultTestBlockRetrieverOptions).(*seekerManager) 355 clockOpts := m.opts.ClockOptions() 356 now := clockOpts.NowFn()() 357 startNano := xtime.ToUnixNano(now) 358 359 fakeTime := now 360 fakeTimeLock := sync.Mutex{} 361 // Setup a function that will allow us to dynamically modify the clock in 362 // a concurrency-safe way 363 newNowFn := func() time.Time { 364 fakeTimeLock.Lock() 365 defer fakeTimeLock.Unlock() 366 return fakeTime 367 } 368 clockOpts = clockOpts.SetNowFn(newNowFn) 369 m.opts = m.opts.SetClockOptions(clockOpts) 370 371 // Initialize some seekers for a time period 372 m.openAnyUnopenSeekersFn = func(byTime *seekersByTime) error { 373 byTime.Lock() 374 defer byTime.Unlock() 375 376 // Don't overwrite if called again 377 if len(byTime.seekers) != 0 { 378 return nil 379 } 380 381 // Don't re-open if they should have expired 382 fakeTimeLock.Lock() 383 defer fakeTimeLock.Unlock() 384 if !fakeTime.Equal(now) { 385 return nil 386 } 387 388 mock := NewMockDataFileSetSeeker(ctrl) 389 mock.EXPECT().Close().Return(nil) 390 mocks := []borrowableSeeker{} 391 mocks = append(mocks, borrowableSeeker{seeker: mock}) 392 byTime.seekers[startNano] = rotatableSeekers{ 393 active: seekersAndBloom{ 394 seekers: mocks, 395 bloomFilter: nil, 396 }, 397 } 398 return nil 399 } 400 401 // Notified everytime the openCloseLoop ticks 402 tickCh := make(chan struct{}) 403 cleanupCh := make(chan struct{}) 404 405 m.sleepFn = func(_ time.Duration) { 406 tickCh <- struct{}{} 407 } 408 409 shards := []uint32{2, 5, 9, 478, 1023} 410 metadata := testNs1Metadata(t) 411 shardSet, err := sharding.NewShardSet( 412 sharding.NewShards(shards, shard.Available), 413 sharding.DefaultHashFn(1), 414 ) 415 require.NoError(t, err) 416 require.NoError(t, m.Open(metadata, shardSet)) 417 418 // Force all the seekers to be opened 419 require.NoError(t, m.CacheShardIndices(shards)) 420 421 seekers := []ConcurrentDataFileSetSeeker{} 422 423 // Steps is a series of steps for the test. It is guaranteed that at least 424 // one (not exactly one!) tick of the openCloseLoop will occur between every step. 425 steps := []struct { 426 title string 427 step func() 428 }{ 429 { 430 title: "Make sure it didn't clean up the seekers which are still in retention", 431 step: func() { 432 m.RLock() 433 for _, shard := range shards { 434 byTime, ok := m.seekersByTime(shard) 435 require.True(t, ok) 436 437 require.Equal(t, 1, len(byTime.seekers[startNano].active.seekers)) 438 } 439 m.RUnlock() 440 }, 441 }, 442 { 443 title: "Borrow a seeker from each shard and then modify the clock such that they're out of retention", 444 step: func() { 445 for _, shard := range shards { 446 seeker, err := m.Borrow(shard, startNano) 447 require.NoError(t, err) 448 require.NotNil(t, seeker) 449 seekers = append(seekers, seeker) 450 } 451 452 fakeTimeLock.Lock() 453 fakeTime = fakeTime.Add(10 * metadata.Options().RetentionOptions().RetentionPeriod()) 454 fakeTimeLock.Unlock() 455 }, 456 }, 457 { 458 title: "Make sure the seeker manager cant be closed while seekers are borrowed", 459 step: func() { 460 require.Equal(t, errCantCloseSeekerManagerWhileSeekersAreBorrowed, m.Close()) 461 }, 462 }, 463 { 464 title: "Make sure that none of the seekers were cleaned up during the openCloseLoop tick (because they're still borrowed)", 465 step: func() { 466 m.RLock() 467 for _, shard := range shards { 468 byTime, ok := m.seekersByTime(shard) 469 require.True(t, ok) 470 require.Equal(t, 1, len(byTime.seekers[startNano].active.seekers)) 471 } 472 m.RUnlock() 473 }, 474 }, 475 { 476 title: "Return the borrowed seekers", 477 step: func() { 478 for i, seeker := range seekers { 479 require.NoError(t, m.Return(shards[i], startNano, seeker)) 480 } 481 }, 482 }, 483 { 484 title: "Make sure that the returned seekers were cleaned up during the openCloseLoop tick", 485 step: func() { 486 m.RLock() 487 for _, shard := range shards { 488 byTime, ok := m.seekersByTime(shard) 489 require.True(t, ok) 490 byTime.RLock() 491 _, ok = byTime.seekers[startNano] 492 byTime.RUnlock() 493 require.False(t, ok) 494 } 495 m.RUnlock() 496 }, 497 }, 498 } 499 500 for _, step := range steps { 501 // Wait for two notifications between steps to guarantee that the entirety 502 // of the openCloseLoop is executed at least once 503 <-tickCh 504 <-tickCh 505 step.step() 506 } 507 508 // Background goroutine that will pull notifications off the tickCh so that 509 // the openCloseLoop is not blocked when we call Close() 510 go func() { 511 for { 512 select { 513 case <-tickCh: 514 continue 515 case <-cleanupCh: 516 return 517 } 518 } 519 }() 520 521 // Restore previous interval once the openCloseLoop ends 522 require.NoError(t, m.Close()) 523 // Make sure there are no goroutines still trying to write into the tickCh 524 // to prevent the test itself from interfering with the goroutine leak test 525 close(cleanupCh) 526 } 527 528 func TestSeekerManagerAssignShardSet(t *testing.T) { 529 defer leaktest.CheckTimeout(t, 1*time.Minute)() 530 531 var ( 532 ctrl = xtest.NewController(t) 533 shards = []uint32{1, 2} 534 m = NewSeekerManager(nil, testDefaultOpts, defaultTestBlockRetrieverOptions).(*seekerManager) 535 ) 536 defer ctrl.Finish() 537 538 var ( 539 wg sync.WaitGroup 540 mockSeekerStatsLock sync.Mutex 541 numMockSeekerClosesByShardAndBlockStart = make(map[uint32]map[xtime.UnixNano]int) 542 ) 543 m.newOpenSeekerFn = func( 544 shard uint32, 545 blockStart xtime.UnixNano, 546 volume int, 547 ) (DataFileSetSeeker, error) { 548 // We expect `defaultTestingFetchConcurrency` number of calls to Close because we return this 549 // many numbers of clones and each clone will need to be closed. 550 wg.Add(defaultTestingFetchConcurrency) 551 552 mock := NewMockDataFileSetSeeker(ctrl) 553 // ConcurrentClone() will be called fetchConcurrency-1 times because the original can be used 554 // as one of the clones. 555 mock.EXPECT().ConcurrentClone().Times(defaultTestingFetchConcurrency-1).Return(mock, nil) 556 mock.EXPECT().Close().Times(defaultTestingFetchConcurrency).DoAndReturn(func() error { 557 mockSeekerStatsLock.Lock() 558 numMockSeekerClosesByBlockStart, ok := numMockSeekerClosesByShardAndBlockStart[shard] 559 if !ok { 560 numMockSeekerClosesByBlockStart = make(map[xtime.UnixNano]int) 561 numMockSeekerClosesByShardAndBlockStart[shard] = numMockSeekerClosesByBlockStart 562 } 563 numMockSeekerClosesByBlockStart[blockStart]++ 564 mockSeekerStatsLock.Unlock() 565 wg.Done() 566 return nil 567 }) 568 mock.EXPECT().ConcurrentIDBloomFilter().Return(nil).AnyTimes() 569 return mock, nil 570 } 571 m.sleepFn = func(_ time.Duration) { 572 time.Sleep(time.Millisecond) 573 } 574 575 metadata := testNs1Metadata(t) 576 shardSet, err := sharding.NewShardSet( 577 sharding.NewShards(shards, shard.Available), 578 sharding.DefaultHashFn(1), 579 ) 580 require.NoError(t, err) 581 // Pick a start time thats within retention so the background loop doesn't close 582 // the seeker. 583 blockStart := xtime.Now().Truncate(metadata.Options().RetentionOptions().BlockSize()) 584 require.NoError(t, m.Open(metadata, shardSet)) 585 586 for _, shard := range shards { 587 seeker, err := m.Borrow(shard, blockStart) 588 require.NoError(t, err) 589 require.NoError(t, m.Return(shard, blockStart, seeker)) 590 } 591 592 // Ensure that UpdateOpenLease() updates the volumes. 593 for _, shard := range shards { 594 updateResult, err := m.UpdateOpenLease(block.LeaseDescriptor{ 595 Namespace: metadata.ID(), 596 Shard: shard, 597 BlockStart: blockStart, 598 }, block.LeaseState{Volume: 1}) 599 require.NoError(t, err) 600 require.Equal(t, block.UpdateOpenLease, updateResult) 601 602 byTime, ok := m.seekersByTime(shard) 603 require.True(t, ok) 604 byTime.RLock() 605 byTime.RUnlock() 606 } 607 608 mockSeekerStatsLock.Lock() 609 for _, numMockSeekerClosesByBlockStart := range numMockSeekerClosesByShardAndBlockStart { 610 require.Equal(t, 611 defaultTestingFetchConcurrency, 612 numMockSeekerClosesByBlockStart[blockStart]) 613 } 614 mockSeekerStatsLock.Unlock() 615 616 // Shards have moved off the node so we assign an empty shard set. 617 m.AssignShardSet(sharding.NewEmptyShardSet(sharding.DefaultHashFn(1))) 618 // Wait until the open/close loop has finished closing all the shards marked to be closed. 619 wg.Wait() 620 621 // Verify that shards are no longer available. 622 for _, shard := range shards { 623 ok, err := m.Test(nil, shard, blockStart) 624 require.Equal(t, errShardNotExists, err) 625 require.False(t, ok) 626 _, err = m.Borrow(shard, blockStart) 627 require.Equal(t, errShardNotExists, err) 628 } 629 630 // Verify that we see the expected # of closes per block start. 631 mockSeekerStatsLock.Lock() 632 for _, numMockSeekerClosesByBlockStart := range numMockSeekerClosesByShardAndBlockStart { 633 for start, numMockSeekerCloses := range numMockSeekerClosesByBlockStart { 634 if blockStart == start { 635 // NB(bodu): These get closed twice since they've been closed once already due to updating their block lease. 636 require.Equal(t, defaultTestingFetchConcurrency*2, numMockSeekerCloses) 637 continue 638 } 639 require.Equal(t, defaultTestingFetchConcurrency, numMockSeekerCloses) 640 } 641 } 642 mockSeekerStatsLock.Unlock() 643 644 // Shards have moved back to the node so we assign a populated shard set again. 645 m.AssignShardSet(shardSet) 646 // Ensure that we can (once again) borrow the shards. 647 for _, shard := range shards { 648 seeker, err := m.Borrow(shard, blockStart) 649 require.NoError(t, err) 650 require.NoError(t, m.Return(shard, blockStart, seeker)) 651 } 652 653 require.NoError(t, m.Close()) 654 } 655 656 // TestSeekerManagerCacheShardIndicesSkipNotFound tests that expired (not found) index filesets 657 // do not return an error. 658 func TestSeekerManagerCacheShardIndicesSkipNotFound(t *testing.T) { 659 defer leaktest.CheckTimeout(t, 1*time.Minute)() 660 661 m := NewSeekerManager(nil, testDefaultOpts, defaultTestBlockRetrieverOptions).(*seekerManager) 662 663 m.newOpenSeekerFn = func( 664 shard uint32, 665 blockStart xtime.UnixNano, 666 volume int, 667 ) (DataFileSetSeeker, error) { 668 return nil, errSeekerManagerFileSetNotFound 669 } 670 671 shards := []uint32{2, 5, 9, 478, 1023} 672 metadata := testNs1Metadata(t) 673 shardSet, err := sharding.NewShardSet( 674 sharding.NewShards(shards, shard.Available), 675 sharding.DefaultHashFn(1), 676 ) 677 require.NoError(t, err) 678 require.NoError(t, m.Open(metadata, shardSet)) 679 680 require.NoError(t, m.CacheShardIndices(shards)) 681 682 require.NoError(t, m.Close()) 683 } 684 685 func TestSeekerManagerDoNotOpenSeekersForOutOfRetentionBlocks(t *testing.T) { 686 defer leaktest.CheckTimeout(t, 1*time.Minute)() 687 var ( 688 ctrl = xtest.NewController(t) 689 shards = []uint32{0} 690 metadata = testNs1Metadata(t) 691 rOpts = metadata.Options().RetentionOptions() 692 blockSize = rOpts.BlockSize() 693 signal = make(chan struct{}) 694 openSeekers = make(map[xtime.UnixNano]struct{}) 695 now = time.Now() 696 opts = NewOptions() 697 ) 698 shardSet, err := sharding.NewShardSet( 699 sharding.NewShards(shards, shard.Available), 700 sharding.DefaultHashFn(1), 701 ) 702 require.NoError(t, err) 703 opts = opts.SetClockOptions(opts.ClockOptions().SetNowFn(func() time.Time { 704 return now 705 })) 706 m := NewSeekerManager(nil, opts, defaultTestBlockRetrieverOptions).(*seekerManager) 707 m.sleepFn = func(_ time.Duration) { 708 signal <- struct{}{} // signal once to indicate that openCloseLoop completed. 709 m.sleepFn = time.Sleep 710 } 711 require.NoError(t, m.Open(metadata, shardSet)) 712 defer func() { 713 require.NoError(t, m.Close()) 714 }() 715 716 m.newOpenSeekerFn = func(shard uint32, blockStart xtime.UnixNano, volume int) (DataFileSetSeeker, error) { 717 openSeekers[blockStart] = struct{}{} 718 mockSeeker := NewMockDataFileSetSeeker(ctrl) 719 mockConcurrentDataFileSetSeeker := NewMockConcurrentDataFileSetSeeker(ctrl) 720 mockConcurrentDataFileSetSeeker.EXPECT().Close().Return(nil) 721 mockSeeker.EXPECT().ConcurrentClone().Return(mockConcurrentDataFileSetSeeker, nil) 722 mockSeeker.EXPECT().ConcurrentIDBloomFilter().Return(nil) 723 mockSeeker.EXPECT().Close().Return(nil) 724 return mockSeeker, nil 725 } 726 727 earliestBlockStart := retention.FlushTimeStart(rOpts, xtime.ToUnixNano(now)) 728 require.NoError(t, m.CacheShardIndices(shards)) 729 730 <-signal 731 require.Contains(t, openSeekers, earliestBlockStart) 732 require.Contains(t, openSeekers, earliestBlockStart.Add(blockSize)) 733 require.NotContains(t, openSeekers, earliestBlockStart.Add(-blockSize)) 734 require.NotContains(t, openSeekers, earliestBlockStart.Add(-2*blockSize)) 735 }