github.com/m3db/m3@v1.5.0/src/dbnode/storage/block/block_test.go (about) 1 // Copyright (c) 2016 Uber Technologies, Inc. 2 // 3 // Permission is hereby granted, free of charge, to any person obtaining a copy 4 // of this software and associated documentation files (the "Software"), to deal 5 // in the Software without restriction, including without limitation the rights 6 // to use, copy, modify, merge, publish, distribute, sublicense, and/or sell 7 // copies of the Software, and to permit persons to whom the Software is 8 // furnished to do so, subject to the following conditions: 9 // 10 // The above copyright notice and this permission notice shall be included in 11 // all copies or substantial portions of the Software. 12 // 13 // THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR 14 // IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, 15 // FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE 16 // AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER 17 // LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, 18 // OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN 19 // THE SOFTWARE. 20 21 package block 22 23 import ( 24 "sync" 25 "testing" 26 "time" 27 28 "github.com/m3db/m3/src/dbnode/encoding" 29 "github.com/m3db/m3/src/dbnode/encoding/m3tsz" 30 "github.com/m3db/m3/src/dbnode/ts" 31 "github.com/m3db/m3/src/dbnode/x/xio" 32 "github.com/m3db/m3/src/x/context" 33 "github.com/m3db/m3/src/x/ident" 34 xtime "github.com/m3db/m3/src/x/time" 35 36 "github.com/golang/mock/gomock" 37 "github.com/m3db/m3/src/dbnode/namespace" 38 "github.com/stretchr/testify/assert" 39 "github.com/stretchr/testify/require" 40 ) 41 42 func testDatabaseBlock(ctrl *gomock.Controller) *dbBlock { 43 opts := NewOptions() 44 b := NewDatabaseBlock(xtime.Now(), 0, ts.Segment{}, opts, namespace.Context{}).(*dbBlock) 45 return b 46 } 47 48 func testDatabaseSeriesBlocks() *databaseSeriesBlocks { 49 return NewDatabaseSeriesBlocks(0).(*databaseSeriesBlocks) 50 } 51 52 func testDatabaseSeriesBlocksWithTimes(times []xtime.UnixNano, sizes []time.Duration) *databaseSeriesBlocks { 53 opts := NewOptions() 54 blocks := testDatabaseSeriesBlocks() 55 for i := range times { 56 block := opts.DatabaseBlockPool().Get() 57 block.Reset(times[i], sizes[i], ts.Segment{}, namespace.Context{}) 58 blocks.AddBlock(block) 59 } 60 return blocks 61 } 62 63 func validateBlocks(t *testing.T, blocks *databaseSeriesBlocks, minTime, 64 maxTime xtime.UnixNano, expectedTimes []xtime.UnixNano, expectedSizes []time.Duration) { 65 require.Equal(t, minTime, blocks.MinTime()) 66 require.Equal(t, maxTime, blocks.MaxTime()) 67 allBlocks := blocks.elems 68 require.Equal(t, len(expectedTimes), len(allBlocks)) 69 for i, timestamp := range expectedTimes { 70 block, exists := allBlocks[timestamp] 71 require.True(t, exists) 72 assert.Equal(t, block.BlockSize(), expectedSizes[i]) 73 } 74 } 75 76 func TestDatabaseBlockReadFromClosedBlock(t *testing.T) { 77 ctrl := gomock.NewController(t) 78 defer ctrl.Finish() 79 80 ctx := context.NewBackground() 81 defer ctx.Close() 82 83 block := testDatabaseBlock(ctrl) 84 block.Close() 85 _, err := block.Stream(ctx) 86 require.Equal(t, errReadFromClosedBlock, err) 87 } 88 89 func TestDatabaseBlockChecksum(t *testing.T) { 90 ctrl := gomock.NewController(t) 91 defer ctrl.Finish() 92 93 block := testDatabaseBlock(ctrl) 94 block.checksum = uint32(10) 95 96 checksum, err := block.Checksum() 97 require.NoError(t, err) 98 require.Equal(t, block.checksum, checksum) 99 } 100 101 type segmentReaderFinalizeCounter struct { 102 xio.SegmentReader 103 // Use a pointer so we can update it from the Finalize method 104 // which must not be a pointer receiver (in order to satisfy 105 // the interface) 106 finalizeCount *int 107 } 108 109 func (r segmentReaderFinalizeCounter) Finalize() { 110 *r.finalizeCount++ 111 } 112 113 // TestDatabaseBlockMerge lazily merges two blocks and verifies that the correct 114 // data is returned, as well as that the underlying streams are not double-finalized 115 // (regression test). 116 func TestDatabaseBlockMerge(t *testing.T) { 117 ctrl := gomock.NewController(t) 118 defer ctrl.Finish() 119 120 // Test data 121 curr := xtime.Now() 122 data := []ts.Datapoint{ 123 ts.Datapoint{ 124 TimestampNanos: curr, 125 Value: 0, 126 }, 127 ts.Datapoint{ 128 TimestampNanos: curr.Add(time.Second), 129 Value: 1, 130 }, 131 } 132 durations := []time.Duration{ 133 time.Minute, 134 time.Hour, 135 } 136 137 // Mock segment reader pool so we count the number of Finalize() calls 138 segmentReaders := []segmentReaderFinalizeCounter{} 139 mockSegmentReaderPool := xio.NewMockSegmentReaderPool(ctrl) 140 getCall := mockSegmentReaderPool.EXPECT().Get() 141 getCall.DoAndReturn(func() xio.SegmentReader { 142 val := 0 143 reader := segmentReaderFinalizeCounter{ 144 xio.NewSegmentReader(ts.Segment{}), 145 &val, 146 } 147 segmentReaders = append(segmentReaders, reader) 148 return reader 149 }).AnyTimes() 150 151 // Setup 152 blockOpts := NewOptions().SetSegmentReaderPool(mockSegmentReaderPool) 153 encodingOpts := encoding.NewOptions() 154 155 // Create the two blocks we plan to merge 156 encoder := m3tsz.NewEncoder(data[0].TimestampNanos, nil, true, encodingOpts) 157 encoder.Encode(data[0], xtime.Second, nil) 158 seg := encoder.Discard() 159 block1 := NewDatabaseBlock(data[0].TimestampNanos, durations[0], seg, 160 blockOpts, namespace.Context{}).(*dbBlock) 161 162 encoder.Reset(data[1].TimestampNanos, 10, nil) 163 encoder.Encode(data[1], xtime.Second, nil) 164 seg = encoder.Discard() 165 block2 := NewDatabaseBlock(data[1].TimestampNanos, durations[1], seg, 166 blockOpts, namespace.Context{}).(*dbBlock) 167 168 // Lazily merge the two blocks 169 block1.Merge(block2) 170 171 // BlockSize should not change 172 require.Equal(t, durations[0], block1.BlockSize()) 173 174 // Try and read the data back and verify it looks good 175 depCtx := block1.opts.ContextPool().Get() 176 stream, err := block1.Stream(depCtx) 177 require.NoError(t, err) 178 seg, err = stream.Segment() 179 require.NoError(t, err) 180 reader := xio.NewSegmentReader(seg) 181 iter := m3tsz.NewReaderIterator(reader, true, encodingOpts) 182 183 i := 0 184 for iter.Next() { 185 dp, _, _ := iter.Current() 186 require.True(t, data[i].Equal(dp)) 187 i++ 188 } 189 require.NoError(t, iter.Err()) 190 191 // Make sure the checksum was updated 192 mergedChecksum, err := block1.Checksum() 193 require.NoError(t, err) 194 require.Equal(t, seg.CalculateChecksum(), mergedChecksum) 195 196 // Make sure each segment reader was only finalized once 197 require.Equal(t, 3, len(segmentReaders)) 198 depCtx.BlockingClose() 199 block1.Close() 200 block2.Close() 201 for _, segmentReader := range segmentReaders { 202 require.Equal(t, 1, *segmentReader.finalizeCount) 203 } 204 } 205 206 // TestDatabaseBlockMergeRace is similar to TestDatabaseBlockMerge, except it 207 // tries to stream the data in multiple go-routines to ensure the merging isn't 208 // racy, this is a regression test for a known issue. 209 func TestDatabaseBlockMergeRace(t *testing.T) { 210 ctrl := gomock.NewController(t) 211 defer ctrl.Finish() 212 213 var ( 214 numRuns = 1000 215 numRoutines = 20 216 ) 217 218 for i := 0; i < numRuns; i++ { 219 func() { 220 // Test data 221 curr := xtime.Now() 222 data := []ts.Datapoint{ 223 ts.Datapoint{ 224 TimestampNanos: curr, 225 Value: 0, 226 }, 227 ts.Datapoint{ 228 TimestampNanos: curr.Add(time.Second), 229 Value: 1, 230 }, 231 } 232 durations := []time.Duration{ 233 time.Minute, 234 time.Hour, 235 } 236 237 // Setup 238 blockOpts := NewOptions() 239 encodingOpts := encoding.NewOptions() 240 241 // Create the two blocks we plan to merge 242 encoder := m3tsz.NewEncoder(data[0].TimestampNanos, nil, true, encodingOpts) 243 encoder.Encode(data[0], xtime.Second, nil) 244 seg := encoder.Discard() 245 block1 := NewDatabaseBlock(data[0].TimestampNanos, durations[0], seg, blockOpts, namespace.Context{}).(*dbBlock) 246 247 encoder.Reset(data[1].TimestampNanos, 10, nil) 248 encoder.Encode(data[1], xtime.Second, nil) 249 seg = encoder.Discard() 250 block2 := NewDatabaseBlock(data[1].TimestampNanos, durations[1], seg, blockOpts, namespace.Context{}).(*dbBlock) 251 252 // Lazily merge the two blocks 253 block1.Merge(block2) 254 255 var wg sync.WaitGroup 256 wg.Add(numRoutines) 257 258 blockFn := func(block *dbBlock) { 259 defer wg.Done() 260 261 depCtx := block.opts.ContextPool().Get() 262 var ( 263 // Make sure we shadow the top level variables 264 // with the same name 265 stream xio.BlockReader 266 seg ts.Segment 267 err error 268 ) 269 stream, err = block.Stream(depCtx) 270 block.Close() 271 if err == errReadFromClosedBlock { 272 return 273 } 274 require.NoError(t, err) 275 276 seg, err = stream.Segment() 277 require.NoError(t, err) 278 reader := xio.NewSegmentReader(seg) 279 iter := m3tsz.NewReaderIterator(reader, true, encodingOpts) 280 281 i := 0 282 for iter.Next() { 283 dp, _, _ := iter.Current() 284 require.True(t, data[i].Equal(dp)) 285 i++ 286 } 287 require.NoError(t, iter.Err()) 288 } 289 290 for i := 0; i < numRoutines; i++ { 291 go blockFn(block1) 292 } 293 294 wg.Wait() 295 }() 296 } 297 } 298 299 // TestDatabaseBlockMergeChained is similar to TestDatabaseBlockMerge except 300 // we try chaining multiple merge calls onto the same block. 301 func TestDatabaseBlockMergeChained(t *testing.T) { 302 ctrl := gomock.NewController(t) 303 defer ctrl.Finish() 304 305 // Test data 306 curr := xtime.Now() 307 data := []ts.Datapoint{ 308 ts.Datapoint{ 309 TimestampNanos: curr, 310 Value: 0, 311 }, 312 ts.Datapoint{ 313 TimestampNanos: curr.Add(time.Second), 314 Value: 1, 315 }, 316 ts.Datapoint{ 317 TimestampNanos: curr.Add(2 * time.Second), 318 Value: 2, 319 }, 320 } 321 durations := []time.Duration{ 322 time.Second, 323 time.Minute, 324 time.Hour, 325 } 326 327 // Mock segment reader pool so we count the number of Finalize() calls 328 segmentReaders := []segmentReaderFinalizeCounter{} 329 mockSegmentReaderPool := xio.NewMockSegmentReaderPool(ctrl) 330 getCall := mockSegmentReaderPool.EXPECT().Get() 331 getCall.DoAndReturn(func() xio.SegmentReader { 332 val := 0 333 reader := segmentReaderFinalizeCounter{ 334 xio.NewSegmentReader(ts.Segment{}), 335 &val, 336 } 337 segmentReaders = append(segmentReaders, reader) 338 return reader 339 }).AnyTimes() 340 341 // Setup 342 blockOpts := NewOptions().SetSegmentReaderPool(mockSegmentReaderPool) 343 encodingOpts := encoding.NewOptions() 344 345 // Create the two blocks we plan to merge 346 encoder := m3tsz.NewEncoder(data[0].TimestampNanos, nil, true, encodingOpts) 347 encoder.Encode(data[0], xtime.Second, nil) 348 seg := encoder.Discard() 349 block1 := NewDatabaseBlock(data[0].TimestampNanos, durations[0], seg, blockOpts, namespace.Context{}).(*dbBlock) 350 351 encoder.Reset(data[1].TimestampNanos, 10, nil) 352 encoder.Encode(data[1], xtime.Second, nil) 353 seg = encoder.Discard() 354 block2 := NewDatabaseBlock(data[1].TimestampNanos, durations[1], seg, blockOpts, namespace.Context{}).(*dbBlock) 355 356 encoder.Reset(data[2].TimestampNanos, 10, nil) 357 encoder.Encode(data[2], xtime.Second, nil) 358 seg = encoder.Discard() 359 block3 := NewDatabaseBlock(data[2].TimestampNanos, durations[2], seg, blockOpts, namespace.Context{}).(*dbBlock) 360 361 // Lazily merge two blocks into block1 362 block1.Merge(block2) 363 block1.Merge(block3) 364 365 // BlockSize should not change 366 require.Equal(t, durations[0], block1.BlockSize()) 367 368 // Try and read the data back and verify it looks good 369 depCtx := block1.opts.ContextPool().Get() 370 stream, err := block1.Stream(depCtx) 371 require.NoError(t, err) 372 seg, err = stream.Segment() 373 require.NoError(t, err) 374 reader := xio.NewSegmentReader(seg) 375 iter := m3tsz.NewReaderIterator(reader, true, encodingOpts) 376 377 i := 0 378 for iter.Next() { 379 dp, _, _ := iter.Current() 380 require.True(t, data[i].Equal(dp)) 381 i++ 382 } 383 require.NoError(t, iter.Err()) 384 385 // Make sure the checksum was updated 386 mergedChecksum, err := block1.Checksum() 387 require.NoError(t, err) 388 require.Equal(t, seg.CalculateChecksum(), mergedChecksum) 389 390 // Make sure each segment reader was only finalized once 391 require.Equal(t, 5, len(segmentReaders)) 392 depCtx.BlockingClose() 393 block1.Close() 394 block2.Close() 395 for _, segmentReader := range segmentReaders { 396 require.Equal(t, 1, *segmentReader.finalizeCount) 397 } 398 } 399 400 func TestDatabaseBlockMergeErrorFromDisk(t *testing.T) { 401 ctrl := gomock.NewController(t) 402 defer ctrl.Finish() 403 404 // Setup 405 var ( 406 curr = xtime.Now() 407 blockOpts = NewOptions() 408 ) 409 410 // Create the two blocks we plan to merge 411 block1 := NewDatabaseBlock(curr, 0, ts.Segment{}, blockOpts, namespace.Context{}).(*dbBlock) 412 block2 := NewDatabaseBlock(curr, 0, ts.Segment{}, blockOpts, namespace.Context{}).(*dbBlock) 413 414 // Mark only block 2 as retrieved from disk so we can make sure it checks 415 // the block that is being merged, as well as the one that is being merged 416 // into. 417 block2.wasRetrievedFromDisk = true 418 419 require.Equal(t, false, block1.wasRetrievedFromDisk) 420 require.Equal(t, errTriedToMergeBlockFromDisk, block1.Merge(block2)) 421 require.Equal(t, errTriedToMergeBlockFromDisk, block2.Merge(block1)) 422 } 423 424 // TestDatabaseBlockChecksumMergesAndRecalculates makes sure that the Checksum method 425 // will check if a lazy-merge is pending, and if so perform it and recalculate the checksum. 426 func TestDatabaseBlockChecksumMergesAndRecalculates(t *testing.T) { 427 ctrl := gomock.NewController(t) 428 defer ctrl.Finish() 429 430 // Test data 431 curr := xtime.Now() 432 data := []ts.Datapoint{ 433 ts.Datapoint{ 434 TimestampNanos: curr, 435 Value: 0, 436 }, 437 ts.Datapoint{ 438 TimestampNanos: curr.Add(time.Second), 439 Value: 1, 440 }, 441 } 442 durations := []time.Duration{ 443 time.Minute, 444 time.Hour, 445 } 446 447 // Setup 448 blockOpts := NewOptions() 449 encodingOpts := encoding.NewOptions() 450 451 // Create the two blocks we plan to merge 452 encoder := m3tsz.NewEncoder(data[0].TimestampNanos, nil, true, encodingOpts) 453 encoder.Encode(data[0], xtime.Second, nil) 454 seg := encoder.Discard() 455 block1 := NewDatabaseBlock(data[0].TimestampNanos, durations[0], seg, blockOpts, namespace.Context{}).(*dbBlock) 456 457 encoder.Reset(data[1].TimestampNanos, 10, nil) 458 encoder.Encode(data[1], xtime.Second, nil) 459 seg = encoder.Discard() 460 block2 := NewDatabaseBlock(data[1].TimestampNanos, durations[1], seg, blockOpts, namespace.Context{}).(*dbBlock) 461 462 // Keep track of the old checksum so we can make sure it changed 463 oldChecksum, err := block1.Checksum() 464 require.NoError(t, err) 465 466 // Lazily merge the two blocks 467 block1.Merge(block2) 468 469 // BlockSize should not change 470 require.Equal(t, durations[0], block1.BlockSize()) 471 472 // Make sure the checksum was updated 473 newChecksum, err := block1.Checksum() 474 require.NoError(t, err) 475 require.NotEqual(t, oldChecksum, newChecksum) 476 477 // Try and read the data back and verify it looks good 478 depCtx := block1.opts.ContextPool().Get() 479 stream, err := block1.Stream(depCtx) 480 require.NoError(t, err) 481 seg, err = stream.Segment() 482 require.NoError(t, err) 483 reader := xio.NewSegmentReader(seg) 484 iter := m3tsz.NewReaderIterator(reader, true, encodingOpts) 485 486 i := 0 487 for iter.Next() { 488 dp, _, _ := iter.Current() 489 require.True(t, data[i].Equal(dp)) 490 i++ 491 } 492 require.NoError(t, iter.Err()) 493 494 // Make sure the new checksum is correct 495 mergedChecksum, err := block1.Checksum() 496 require.NoError(t, err) 497 require.Equal(t, seg.CalculateChecksum(), mergedChecksum) 498 } 499 500 func TestDatabaseBlockStreamMergePerformsCopy(t *testing.T) { 501 ctrl := gomock.NewController(t) 502 defer ctrl.Finish() 503 504 // Test data 505 curr := xtime.Now() 506 data := []ts.Datapoint{ 507 ts.Datapoint{ 508 TimestampNanos: curr, 509 Value: 0, 510 }, 511 ts.Datapoint{ 512 TimestampNanos: curr.Add(time.Second), 513 Value: 1, 514 }, 515 } 516 durations := []time.Duration{ 517 time.Minute, 518 time.Hour, 519 } 520 521 // Setup 522 blockOpts := NewOptions() 523 encodingOpts := encoding.NewOptions() 524 525 // Create the two blocks we plan to merge 526 encoder := m3tsz.NewEncoder(data[0].TimestampNanos, nil, true, encodingOpts) 527 encoder.Encode(data[0], xtime.Second, nil) 528 seg := encoder.Discard() 529 block1 := NewDatabaseBlock(data[0].TimestampNanos, durations[0], seg, blockOpts, namespace.Context{}).(*dbBlock) 530 531 encoder.Reset(data[1].TimestampNanos, 10, nil) 532 encoder.Encode(data[1], xtime.Second, nil) 533 seg = encoder.Discard() 534 block2 := NewDatabaseBlock(data[1].TimestampNanos, durations[1], seg, blockOpts, namespace.Context{}).(*dbBlock) 535 536 err := block1.Merge(block2) 537 require.NoError(t, err) 538 539 depCtx := block1.opts.ContextPool().Get() 540 stream, err := block1.Stream(depCtx) 541 require.NoError(t, err) 542 block1.Close() 543 544 seg, err = stream.Segment() 545 require.NoError(t, err) 546 reader := xio.NewSegmentReader(seg) 547 iter := m3tsz.NewReaderIterator(reader, true, encodingOpts) 548 549 i := 0 550 for iter.Next() { 551 dp, _, _ := iter.Current() 552 require.True(t, data[i].Equal(dp)) 553 i++ 554 } 555 require.NoError(t, iter.Err()) 556 } 557 558 func TestDatabaseBlockCloseIfFromDisk(t *testing.T) { 559 var ( 560 blockOpts = NewOptions() 561 blockNotFromDisk = NewDatabaseBlock(0, time.Hour, ts.Segment{}, blockOpts, namespace.Context{}).(*dbBlock) 562 blockFromDisk = NewDatabaseBlock(0, time.Hour, ts.Segment{}, blockOpts, namespace.Context{}).(*dbBlock) 563 ) 564 blockFromDisk.wasRetrievedFromDisk = true 565 566 require.False(t, blockNotFromDisk.CloseIfFromDisk()) 567 require.True(t, blockFromDisk.CloseIfFromDisk()) 568 } 569 570 func TestDatabaseSeriesBlocksAddBlock(t *testing.T) { 571 now := xtime.Now() 572 blockTimes := []xtime.UnixNano{ 573 now, now.Add(time.Second), 574 now.Add(time.Minute), now.Add(-time.Second), now.Add(-time.Hour), 575 } 576 blockSizes := []time.Duration{ 577 time.Minute, time.Hour, time.Second, 578 time.Microsecond, time.Millisecond, 579 } 580 blocks := testDatabaseSeriesBlocksWithTimes(blockTimes, blockSizes) 581 validateBlocks(t, blocks, blockTimes[4], blockTimes[2], blockTimes, blockSizes) 582 } 583 584 func TestDatabaseSeriesBlocksAddSeries(t *testing.T) { 585 now := xtime.Now() 586 blockTimes := [][]xtime.UnixNano{ 587 {now, now.Add(time.Second), now.Add(time.Minute), now.Add(-time.Second), now.Add(-time.Hour)}, 588 {now.Add(-time.Minute), now.Add(time.Hour)}, 589 } 590 blockSizes := [][]time.Duration{ 591 {time.Minute, time.Hour, time.Second, time.Microsecond, time.Millisecond}, 592 {time.Minute * 2, time.Hour * 21}, 593 } 594 blocks := testDatabaseSeriesBlocksWithTimes(blockTimes[0], blockSizes[0]) 595 other := testDatabaseSeriesBlocksWithTimes(blockTimes[1], blockSizes[1]) 596 blocks.AddSeries(other) 597 var expectedTimes []xtime.UnixNano 598 for _, bt := range blockTimes { 599 expectedTimes = append(expectedTimes, bt...) 600 } 601 var expectedSizes []time.Duration 602 for _, bt := range blockSizes { 603 expectedSizes = append(expectedSizes, bt...) 604 } 605 606 validateBlocks(t, blocks, expectedTimes[4], expectedTimes[6], expectedTimes, expectedSizes) 607 } 608 609 func TestDatabaseSeriesBlocksGetBlockAt(t *testing.T) { 610 now := xtime.Now() 611 blockTimes := []xtime.UnixNano{now, now.Add(time.Second), now.Add(-time.Hour)} 612 blockSizes := []time.Duration{time.Minute, time.Hour, time.Second} 613 614 blocks := testDatabaseSeriesBlocksWithTimes(blockTimes, blockSizes) 615 for i, bt := range blockTimes { 616 b, exists := blocks.BlockAt(bt) 617 require.True(t, exists) 618 require.Equal(t, b.BlockSize(), blockSizes[i]) 619 } 620 _, exists := blocks.BlockAt(now.Add(time.Minute)) 621 require.False(t, exists) 622 } 623 624 func TestDatabaseSeriesBlocksRemoveBlockAt(t *testing.T) { 625 now := xtime.Now() 626 blockTimes := []xtime.UnixNano{now, now.Add(-time.Second), now.Add(time.Hour)} 627 blockSizes := []time.Duration{time.Minute, time.Hour, time.Second} 628 629 blocks := testDatabaseSeriesBlocksWithTimes(blockTimes, blockSizes) 630 blocks.RemoveBlockAt(now.Add(-time.Hour)) 631 validateBlocks(t, blocks, blockTimes[1], blockTimes[2], blockTimes, blockSizes) 632 633 expected := []struct { 634 min xtime.UnixNano 635 max xtime.UnixNano 636 allTimes []xtime.UnixNano 637 }{ 638 {blockTimes[1], blockTimes[2], blockTimes[1:]}, 639 {blockTimes[2], blockTimes[2], blockTimes[2:]}, 640 {timeZero, timeZero, []xtime.UnixNano{}}, 641 } 642 for i, bt := range blockTimes { 643 blocks.RemoveBlockAt(bt) 644 blockSizes = blockSizes[1:] 645 validateBlocks(t, blocks, expected[i].min, expected[i].max, expected[i].allTimes, blockSizes) 646 } 647 } 648 649 func TestDatabaseSeriesBlocksRemoveAll(t *testing.T) { 650 now := xtime.Now() 651 blockTimes := []xtime.UnixNano{now, now.Add(-time.Second), now.Add(time.Hour)} 652 blockSizes := []time.Duration{time.Minute, time.Hour, time.Second} 653 654 blocks := testDatabaseSeriesBlocksWithTimes(blockTimes, blockSizes) 655 require.Equal(t, len(blockTimes), len(blocks.AllBlocks())) 656 657 blocks.RemoveAll() 658 require.Equal(t, 0, len(blocks.AllBlocks())) 659 } 660 661 func TestDatabaseSeriesBlocksClose(t *testing.T) { 662 now := xtime.Now() 663 blockTimes := []xtime.UnixNano{now, now.Add(-time.Second), now.Add(time.Hour)} 664 blockSizes := []time.Duration{time.Minute, time.Hour, time.Second} 665 666 blocks := testDatabaseSeriesBlocksWithTimes(blockTimes, blockSizes) 667 require.Equal(t, len(blockTimes), len(blocks.AllBlocks())) 668 669 blocks.Close() 670 require.Equal(t, 0, len(blocks.AllBlocks())) 671 672 var nilPointer map[xtime.UnixNano]DatabaseBlock 673 require.Equal(t, nilPointer, blocks.elems) 674 } 675 676 func TestDatabaseSeriesBlocksReset(t *testing.T) { 677 now := xtime.Now() 678 blockTimes := []xtime.UnixNano{now, now.Add(-time.Second), now.Add(time.Hour)} 679 blockSizes := []time.Duration{time.Minute, time.Hour, time.Second} 680 681 blocks := testDatabaseSeriesBlocksWithTimes(blockTimes, blockSizes) 682 require.Equal(t, len(blockTimes), len(blocks.AllBlocks())) 683 684 blocks.Reset() 685 686 require.Equal(t, 0, len(blocks.AllBlocks())) 687 require.Equal(t, 0, len(blocks.elems)) 688 require.True(t, blocks.min.Equal(0)) 689 require.True(t, blocks.max.Equal(0)) 690 } 691 692 func TestBlockResetFromDisk(t *testing.T) { 693 ctrl := gomock.NewController(t) 694 defer ctrl.Finish() 695 696 bl := testDatabaseBlock(ctrl) 697 now := xtime.Now() 698 blockSize := 2 * time.Hour 699 id := ident.StringID("testID") 700 segment := ts.Segment{} 701 bl.ResetFromDisk(now, blockSize, segment, id, namespace.Context{}) 702 703 assert.True(t, now.Equal(bl.StartTime())) 704 assert.Equal(t, blockSize, bl.BlockSize()) 705 assert.Equal(t, segment, bl.segment) 706 assert.Equal(t, id, bl.seriesID) 707 assert.True(t, bl.WasRetrievedFromDisk()) 708 }