github.com/cockroachdb/pebble@v1.1.2/metamorphic/key_manager_test.go (about) 1 package metamorphic 2 3 import ( 4 "bytes" 5 "testing" 6 7 "github.com/cockroachdb/pebble/internal/randvar" 8 "github.com/stretchr/testify/require" 9 ) 10 11 func TestObjKey(t *testing.T) { 12 testCases := []struct { 13 key objKey 14 want string 15 }{ 16 { 17 key: makeObjKey(makeObjID(dbTag, 0), []byte("foo")), 18 want: "db:foo", 19 }, 20 { 21 key: makeObjKey(makeObjID(batchTag, 1), []byte("bar")), 22 want: "batch1:bar", 23 }, 24 } 25 26 for _, tc := range testCases { 27 t.Run("", func(t *testing.T) { 28 require.Equal(t, tc.want, tc.key.String()) 29 }) 30 } 31 } 32 33 func TestGlobalStateIndicatesEligibleForSingleDelete(t *testing.T) { 34 key := makeObjKey(makeObjID(dbTag, 0), []byte("foo")) 35 testCases := []struct { 36 meta keyMeta 37 want bool 38 }{ 39 { 40 meta: keyMeta{ 41 objKey: key, 42 }, 43 want: false, 44 }, 45 { 46 meta: keyMeta{ 47 objKey: key, 48 sets: 1, 49 }, 50 want: true, 51 }, 52 { 53 meta: keyMeta{ 54 objKey: key, 55 sets: 2, 56 }, 57 want: false, 58 }, 59 { 60 meta: keyMeta{ 61 objKey: key, 62 sets: 1, 63 merges: 1, 64 }, 65 want: false, 66 }, 67 { 68 meta: keyMeta{ 69 objKey: key, 70 sets: 1, 71 dels: 1, 72 }, 73 want: false, 74 }, 75 { 76 meta: keyMeta{ 77 objKey: key, 78 sets: 1, 79 singleDel: true, 80 }, 81 want: false, 82 }, 83 } 84 85 for _, tc := range testCases { 86 k := newKeyManager() 87 t.Run("", func(t *testing.T) { 88 k.globalKeysMap[string(key.key)] = &tc.meta 89 require.Equal(t, tc.want, k.globalStateIndicatesEligibleForSingleDelete(key.key)) 90 }) 91 } 92 } 93 94 func TestKeyMeta_MergeInto(t *testing.T) { 95 testCases := []struct { 96 existing keyMeta 97 toMerge keyMeta 98 expected keyMeta 99 }{ 100 { 101 existing: keyMeta{ 102 sets: 1, 103 merges: 0, 104 singleDel: false, 105 }, 106 toMerge: keyMeta{ 107 sets: 0, 108 merges: 0, 109 singleDel: true, 110 }, 111 expected: keyMeta{ 112 sets: 1, 113 merges: 0, 114 singleDel: true, 115 updateOps: []keyUpdate{ 116 {deleted: true, metaTimestamp: 0}, 117 }, 118 }, 119 }, 120 { 121 existing: keyMeta{ 122 sets: 3, 123 merges: 1, 124 dels: 7, 125 }, 126 toMerge: keyMeta{ 127 sets: 4, 128 merges: 2, 129 dels: 8, 130 del: true, 131 }, 132 expected: keyMeta{ 133 sets: 7, 134 merges: 3, 135 dels: 15, 136 del: true, 137 updateOps: []keyUpdate{ 138 {deleted: true, metaTimestamp: 1}, 139 }, 140 }, 141 }, 142 { 143 existing: keyMeta{ 144 sets: 3, 145 merges: 1, 146 dels: 7, 147 del: true, 148 }, 149 toMerge: keyMeta{ 150 sets: 1, 151 merges: 0, 152 dels: 8, 153 del: false, 154 }, 155 expected: keyMeta{ 156 sets: 4, 157 merges: 1, 158 dels: 15, 159 del: false, 160 updateOps: []keyUpdate{ 161 {deleted: false, metaTimestamp: 2}, 162 }, 163 }, 164 }, 165 } 166 167 keyManager := newKeyManager() 168 for _, tc := range testCases { 169 t.Run("", func(t *testing.T) { 170 tc.toMerge.mergeInto(keyManager, &tc.existing) 171 require.Equal(t, tc.expected, tc.existing) 172 }) 173 } 174 } 175 176 func TestKeyManager_AddKey(t *testing.T) { 177 m := newKeyManager() 178 require.Empty(t, m.globalKeys) 179 180 k1 := []byte("foo") 181 require.True(t, m.addNewKey(k1)) 182 require.Len(t, m.globalKeys, 1) 183 require.Len(t, m.globalKeyPrefixes, 1) 184 require.Contains(t, m.globalKeys, k1) 185 require.Contains(t, m.globalKeyPrefixes, k1) 186 require.False(t, m.addNewKey(k1)) 187 require.True(t, m.prefixExists([]byte("foo"))) 188 require.False(t, m.prefixExists([]byte("bar"))) 189 190 k2 := []byte("bar") 191 require.True(t, m.addNewKey(k2)) 192 require.Len(t, m.globalKeys, 2) 193 require.Len(t, m.globalKeyPrefixes, 2) 194 require.Contains(t, m.globalKeys, k2) 195 require.Contains(t, m.globalKeyPrefixes, k2) 196 require.True(t, m.prefixExists([]byte("bar"))) 197 k3 := []byte("bax@4") 198 require.True(t, m.addNewKey(k3)) 199 require.Len(t, m.globalKeys, 3) 200 require.Len(t, m.globalKeyPrefixes, 3) 201 require.Contains(t, m.globalKeys, k3) 202 require.Contains(t, m.globalKeyPrefixes, []byte("bax")) 203 require.True(t, m.prefixExists([]byte("bax"))) 204 k4 := []byte("foo@6") 205 require.True(t, m.addNewKey(k4)) 206 require.Len(t, m.globalKeys, 4) 207 require.Len(t, m.globalKeyPrefixes, 3) 208 require.Contains(t, m.globalKeys, k4) 209 require.True(t, m.prefixExists([]byte("foo"))) 210 211 require.Equal(t, [][]byte{ 212 []byte("bar"), []byte("bax"), []byte("foo"), 213 }, m.prefixes()) 214 } 215 216 func TestKeyManager_GetOrInit(t *testing.T) { 217 id := makeObjID(batchTag, 1) 218 key := []byte("foo") 219 o := makeObjKey(id, key) 220 221 m := newKeyManager() 222 require.NotContains(t, m.byObjKey, o.String()) 223 require.NotContains(t, m.byObj, id) 224 require.Contains(t, m.byObj, makeObjID(dbTag, 0)) // Always contains the DB key. 225 226 meta1 := m.getOrInit(id, key) 227 require.Contains(t, m.byObjKey, o.String()) 228 require.Contains(t, m.byObj, id) 229 230 // Idempotent. 231 meta2 := m.getOrInit(id, key) 232 require.Equal(t, meta1, meta2) 233 } 234 235 func TestKeyManager_Contains(t *testing.T) { 236 id := makeObjID(dbTag, 0) 237 key := []byte("foo") 238 239 m := newKeyManager() 240 require.False(t, m.contains(id, key)) 241 242 m.getOrInit(id, key) 243 require.True(t, m.contains(id, key)) 244 } 245 246 func TestKeyManager_MergeInto(t *testing.T) { 247 fromID := makeObjID(batchTag, 1) 248 toID := makeObjID(dbTag, 0) 249 250 m := newKeyManager() 251 252 // Two keys in "from". 253 a := m.getOrInit(fromID, []byte("foo")) 254 a.sets = 1 255 b := m.getOrInit(fromID, []byte("bar")) 256 b.merges = 2 257 258 // One key in "to", with same value as a key in "from", that will be merged. 259 m.getOrInit(toID, []byte("foo")) 260 261 // Before, there are two sets. 262 require.Len(t, m.byObj[fromID], 2) 263 require.Len(t, m.byObj[toID], 1) 264 265 m.mergeKeysInto(fromID, toID) 266 267 // Keys in "from" sets are moved to "to" set. 268 require.Len(t, m.byObj[toID], 2) 269 270 // Key "foo" was merged into "to". 271 foo := m.getOrInit(toID, []byte("foo")) 272 require.Equal(t, 1, foo.sets) // value was merged. 273 274 // Key "bar" was merged into "to". 275 bar := m.getOrInit(toID, []byte("bar")) 276 require.Equal(t, 2, bar.merges) // value was unchanged. 277 278 // Keys in "from" sets are removed from maps. 279 require.NotContains(t, m.byObjKey, makeObjKey(fromID, a.key)) 280 require.NotContains(t, m.byObjKey, makeObjKey(fromID, b.key)) 281 require.NotContains(t, m.byObj, fromID) 282 } 283 284 type seqFn func(t *testing.T, k *keyManager) 285 286 func updateForOp(op op) seqFn { 287 return func(t *testing.T, k *keyManager) { 288 k.update(op) 289 } 290 } 291 292 func addKey(key []byte, expected bool) seqFn { 293 return func(t *testing.T, k *keyManager) { 294 require.Equal(t, expected, k.addNewKey(key)) 295 } 296 } 297 298 func eligibleRead(key []byte, val bool) seqFn { 299 return func(t *testing.T, k *keyManager) { 300 require.Equal(t, val, contains(key, k.eligibleReadKeys())) 301 } 302 } 303 304 func eligibleWrite(key []byte, val bool) seqFn { 305 return func(t *testing.T, k *keyManager) { 306 require.Equal(t, val, contains(key, k.eligibleWriteKeys())) 307 } 308 } 309 310 func eligibleSingleDelete(key []byte, val bool, id objID) seqFn { 311 return func(t *testing.T, k *keyManager) { 312 require.Equal(t, val, contains(key, k.eligibleSingleDeleteKeys(id))) 313 } 314 } 315 316 func contains(key []byte, keys [][]byte) bool { 317 for _, k := range keys { 318 if bytes.Equal(key, k) { 319 return true 320 } 321 } 322 return false 323 } 324 325 func TestKeyManager(t *testing.T) { 326 var ( 327 id1 = makeObjID(batchTag, 0) 328 id2 = makeObjID(batchTag, 1) 329 key1 = []byte("foo") 330 ) 331 332 testCases := []struct { 333 description string 334 ops []seqFn 335 wantPanic bool 336 }{ 337 { 338 description: "set, single del, on db", 339 ops: []seqFn{ 340 addKey(key1, true), 341 addKey(key1, false), 342 eligibleRead(key1, true), 343 eligibleWrite(key1, true), 344 eligibleSingleDelete(key1, false, dbObjID), 345 eligibleSingleDelete(key1, false, id1), 346 updateForOp(&setOp{writerID: dbObjID, key: key1}), 347 eligibleRead(key1, true), 348 eligibleWrite(key1, true), 349 eligibleSingleDelete(key1, true, dbObjID), 350 eligibleSingleDelete(key1, true, id1), 351 updateForOp(&singleDeleteOp{writerID: dbObjID, key: key1}), 352 eligibleRead(key1, true), 353 eligibleWrite(key1, true), 354 eligibleSingleDelete(key1, false, dbObjID), 355 }, 356 }, 357 { 358 description: "set, single del, on batch", 359 ops: []seqFn{ 360 addKey(key1, true), 361 updateForOp(&setOp{writerID: id1, key: key1}), 362 eligibleRead(key1, true), 363 eligibleWrite(key1, true), 364 eligibleSingleDelete(key1, false, dbObjID), 365 eligibleSingleDelete(key1, true, id1), 366 eligibleSingleDelete(key1, false, id2), 367 updateForOp(&singleDeleteOp{writerID: id1, key: key1}), 368 eligibleRead(key1, true), 369 eligibleWrite(key1, false), 370 eligibleSingleDelete(key1, false, dbObjID), 371 eligibleSingleDelete(key1, false, id1), 372 updateForOp(&applyOp{batchID: id1, writerID: dbObjID}), 373 eligibleWrite(key1, true), 374 eligibleSingleDelete(key1, false, dbObjID), 375 }, 376 }, 377 { 378 description: "set on db, single del on batch", 379 ops: []seqFn{ 380 addKey(key1, true), 381 updateForOp(&setOp{writerID: dbObjID, key: key1}), 382 eligibleWrite(key1, true), 383 eligibleSingleDelete(key1, true, dbObjID), 384 eligibleSingleDelete(key1, true, id1), 385 updateForOp(&singleDeleteOp{writerID: id1, key: key1}), 386 eligibleWrite(key1, false), 387 eligibleSingleDelete(key1, false, dbObjID), 388 eligibleSingleDelete(key1, false, id1), 389 updateForOp(&applyOp{batchID: id1, writerID: dbObjID}), 390 eligibleWrite(key1, true), 391 eligibleSingleDelete(key1, false, dbObjID), 392 updateForOp(&setOp{writerID: dbObjID, key: key1}), 393 eligibleSingleDelete(key1, true, dbObjID), 394 eligibleSingleDelete(key1, true, id1), 395 }, 396 }, 397 { 398 description: "set, del, set, single del, on db", 399 ops: []seqFn{ 400 addKey(key1, true), 401 updateForOp(&setOp{writerID: dbObjID, key: key1}), 402 updateForOp(&deleteOp{writerID: dbObjID, key: key1}), 403 eligibleWrite(key1, true), 404 eligibleSingleDelete(key1, false, dbObjID), 405 updateForOp(&setOp{writerID: dbObjID, key: key1}), 406 eligibleWrite(key1, true), 407 eligibleSingleDelete(key1, true, dbObjID), 408 updateForOp(&singleDeleteOp{writerID: dbObjID, key: key1}), 409 eligibleWrite(key1, true), 410 eligibleSingleDelete(key1, false, dbObjID), 411 }, 412 }, 413 { 414 description: "set, del, set, del, on batches", 415 ops: []seqFn{ 416 addKey(key1, true), 417 updateForOp(&setOp{writerID: id1, key: key1}), 418 updateForOp(&deleteOp{writerID: id1, key: key1}), 419 updateForOp(&setOp{writerID: id1, key: key1}), 420 eligibleWrite(key1, true), 421 eligibleSingleDelete(key1, false, id1), 422 updateForOp(&applyOp{batchID: id1, writerID: dbObjID}), 423 eligibleWrite(key1, true), 424 // Not eligible for single del since the set count is 2. 425 eligibleSingleDelete(key1, false, dbObjID), 426 updateForOp(&setOp{writerID: dbObjID, key: key1}), 427 // Not eligible for single del since the set count is 3. 428 eligibleSingleDelete(key1, false, dbObjID), 429 updateForOp(&deleteOp{writerID: id2, key: key1}), 430 updateForOp(&applyOp{batchID: id2, writerID: dbObjID}), 431 // Set count is 0. 432 eligibleSingleDelete(key1, false, dbObjID), 433 // Set count is 1. 434 updateForOp(&setOp{writerID: dbObjID, key: key1}), 435 eligibleSingleDelete(key1, true, dbObjID), 436 }, 437 }, 438 { 439 description: "set, merge, del, set, single del, on db", 440 ops: []seqFn{ 441 addKey(key1, true), 442 updateForOp(&setOp{writerID: dbObjID, key: key1}), 443 eligibleSingleDelete(key1, true, dbObjID), 444 updateForOp(&mergeOp{writerID: dbObjID, key: key1}), 445 eligibleSingleDelete(key1, false, dbObjID), 446 updateForOp(&deleteOp{writerID: dbObjID, key: key1}), 447 eligibleWrite(key1, true), 448 eligibleSingleDelete(key1, false, dbObjID), 449 updateForOp(&setOp{writerID: dbObjID, key: key1}), 450 eligibleWrite(key1, true), 451 eligibleSingleDelete(key1, true, dbObjID), 452 updateForOp(&singleDeleteOp{writerID: dbObjID, key: key1}), 453 eligibleWrite(key1, true), 454 eligibleSingleDelete(key1, false, dbObjID), 455 }, 456 }, 457 { 458 description: "set, del on db, set, single del on batch", 459 ops: []seqFn{ 460 addKey(key1, true), 461 updateForOp(&setOp{writerID: dbObjID, key: key1}), 462 eligibleSingleDelete(key1, true, dbObjID), 463 updateForOp(&deleteOp{writerID: dbObjID, key: key1}), 464 eligibleWrite(key1, true), 465 eligibleSingleDelete(key1, false, dbObjID), 466 updateForOp(&setOp{writerID: id1, key: key1}), 467 eligibleWrite(key1, true), 468 eligibleSingleDelete(key1, false, dbObjID), 469 eligibleSingleDelete(key1, true, id1), 470 updateForOp(&singleDeleteOp{writerID: id1, key: key1}), 471 eligibleWrite(key1, false), 472 eligibleSingleDelete(key1, false, id1), 473 eligibleSingleDelete(key1, false, dbObjID), 474 updateForOp(&applyOp{batchID: id1, writerID: dbObjID}), 475 eligibleWrite(key1, true), 476 eligibleSingleDelete(key1, false, dbObjID), 477 updateForOp(&setOp{writerID: dbObjID, key: key1}), 478 eligibleSingleDelete(key1, true, dbObjID), 479 }, 480 }, 481 } 482 483 for _, tc := range testCases { 484 t.Run(tc.description, func(t *testing.T) { 485 m := newKeyManager() 486 tFunc := func() { 487 for _, op := range tc.ops { 488 op(t, m) 489 } 490 } 491 if tc.wantPanic { 492 require.Panics(t, tFunc) 493 } else { 494 tFunc() 495 } 496 }) 497 } 498 } 499 500 func TestOpWrittenKeys(t *testing.T) { 501 for name, info := range methods { 502 t.Run(name, func(t *testing.T) { 503 // Any operations that exist in methods but are not handled in 504 // opWrittenKeys will result in a panic, failing the subtest. 505 opWrittenKeys(info.constructor()) 506 }) 507 } 508 } 509 510 func TestLoadPrecedingKeys(t *testing.T) { 511 rng := randvar.NewRand() 512 cfg := defaultConfig() 513 km := newKeyManager() 514 ops := generate(rng, 1000, cfg, km) 515 516 cfg2 := defaultConfig() 517 km2 := newKeyManager() 518 loadPrecedingKeys(t, ops, &cfg2, km2) 519 520 // NB: We can't assert equality, because the original run may not have 521 // ever used the max of the distribution. 522 require.Greater(t, cfg2.writeSuffixDist.Max(), uint64(1)) 523 524 // NB: We can't assert equality, because the original run may have generated 525 // keys that it didn't end up using in operations. 526 require.Subset(t, km.globalKeys, km2.globalKeys) 527 require.Subset(t, km.globalKeyPrefixes, km2.globalKeyPrefixes) 528 }