github.com/cockroachdb/pebble@v1.1.2/compaction_picker_test.go (about)

     1  // Copyright 2018 The LevelDB-Go and Pebble Authors. All rights reserved. Use
     2  // of this source code is governed by a BSD-style license that can be found in
     3  // the LICENSE file.
     4  
     5  package pebble
     6  
     7  import (
     8  	"bytes"
     9  	"fmt"
    10  	"math"
    11  	"sort"
    12  	"strconv"
    13  	"strings"
    14  	"sync"
    15  	"testing"
    16  	"time"
    17  
    18  	"github.com/cockroachdb/datadriven"
    19  	"github.com/cockroachdb/errors"
    20  	"github.com/cockroachdb/pebble/internal/base"
    21  	"github.com/cockroachdb/pebble/internal/humanize"
    22  	"github.com/cockroachdb/pebble/internal/manifest"
    23  	"github.com/cockroachdb/pebble/internal/testkeys"
    24  	"github.com/cockroachdb/pebble/vfs"
    25  	"github.com/stretchr/testify/require"
    26  )
    27  
    28  func loadVersion(t *testing.T, d *datadriven.TestData) (*version, *Options, string) {
    29  	var sizes [numLevels]int64
    30  	opts := &Options{}
    31  	opts.testingRandomized(t)
    32  	opts.EnsureDefaults()
    33  
    34  	if len(d.CmdArgs) != 1 {
    35  		return nil, nil, fmt.Sprintf("%s expects 1 argument", d.Cmd)
    36  	}
    37  	var err error
    38  	opts.LBaseMaxBytes, err = strconv.ParseInt(d.CmdArgs[0].Key, 10, 64)
    39  	if err != nil {
    40  		return nil, nil, err.Error()
    41  	}
    42  
    43  	var files [numLevels][]*fileMetadata
    44  	if len(d.Input) > 0 {
    45  		// Parse each line as
    46  		//
    47  		// <level>: <size> [compensation]
    48  		//
    49  		// Creating sstables within the level whose file sizes total to `size`
    50  		// and whose compensated file sizes total to `size`+`compensation`. If
    51  		// size is sufficiently large, only one single file is created. See
    52  		// the TODO below.
    53  		for _, data := range strings.Split(d.Input, "\n") {
    54  			parts := strings.Split(data, " ")
    55  			parts[0] = strings.TrimSuffix(strings.TrimSpace(parts[0]), ":")
    56  			if len(parts) < 2 {
    57  				return nil, nil, fmt.Sprintf("malformed test:\n%s", d.Input)
    58  			}
    59  			level, err := strconv.Atoi(parts[0])
    60  			if err != nil {
    61  				return nil, nil, err.Error()
    62  			}
    63  			if files[level] != nil {
    64  				return nil, nil, fmt.Sprintf("level %d already filled", level)
    65  			}
    66  			size, err := strconv.ParseUint(strings.TrimSpace(parts[1]), 10, 64)
    67  			if err != nil {
    68  				return nil, nil, err.Error()
    69  			}
    70  			var compensation uint64
    71  			if len(parts) == 3 {
    72  				compensation, err = strconv.ParseUint(strings.TrimSpace(parts[2]), 10, 64)
    73  				if err != nil {
    74  					return nil, nil, err.Error()
    75  				}
    76  			}
    77  
    78  			var lastFile *fileMetadata
    79  			for i := uint64(1); sizes[level] < int64(size); i++ {
    80  				var key InternalKey
    81  				if level == 0 {
    82  					// For L0, make `size` overlapping files.
    83  					key = base.MakeInternalKey([]byte(fmt.Sprintf("%04d", 1)), i, InternalKeyKindSet)
    84  				} else {
    85  					key = base.MakeInternalKey([]byte(fmt.Sprintf("%04d", i)), i, InternalKeyKindSet)
    86  				}
    87  				m := (&fileMetadata{
    88  					FileNum:        base.FileNum(uint64(level)*100_000 + i),
    89  					SmallestSeqNum: key.SeqNum(),
    90  					LargestSeqNum:  key.SeqNum(),
    91  					Size:           1,
    92  					Stats: manifest.TableStats{
    93  						RangeDeletionsBytesEstimate: 0,
    94  					},
    95  				}).ExtendPointKeyBounds(opts.Comparer.Compare, key, key)
    96  				m.InitPhysicalBacking()
    97  				m.StatsMarkValid()
    98  				lastFile = m
    99  				if size >= 100 {
   100  					// If the requested size of the level is very large only add a single
   101  					// file in order to avoid massive blow-up in the number of files in
   102  					// the Version.
   103  					//
   104  					// TODO(peter): There is tension between the testing in
   105  					// TestCompactionPickerLevelMaxBytes and
   106  					// TestCompactionPickerTargetLevel. Clean this up somehow.
   107  					m.Size = size
   108  					if level != 0 {
   109  						endKey := base.MakeInternalKey([]byte(fmt.Sprintf("%04d", size)), i, InternalKeyKindSet)
   110  						m.ExtendPointKeyBounds(opts.Comparer.Compare, key, endKey)
   111  					}
   112  				}
   113  				files[level] = append(files[level], m)
   114  				sizes[level] += int64(m.Size)
   115  			}
   116  			// Let all the compensation be due to the last file.
   117  			if lastFile != nil && compensation > 0 {
   118  				lastFile.Stats.RangeDeletionsBytesEstimate = compensation
   119  			}
   120  		}
   121  	}
   122  
   123  	vers := newVersion(opts, files)
   124  	return vers, opts, ""
   125  }
   126  
   127  func TestCompactionPickerByScoreLevelMaxBytes(t *testing.T) {
   128  	datadriven.RunTest(t, "testdata/compaction_picker_level_max_bytes",
   129  		func(t *testing.T, d *datadriven.TestData) string {
   130  			switch d.Cmd {
   131  			case "init":
   132  				vers, opts, errMsg := loadVersion(t, d)
   133  				if errMsg != "" {
   134  					return errMsg
   135  				}
   136  
   137  				p, ok := newCompactionPicker(vers, opts, nil).(*compactionPickerByScore)
   138  				require.True(t, ok)
   139  				var buf bytes.Buffer
   140  				for level := p.getBaseLevel(); level < numLevels; level++ {
   141  					fmt.Fprintf(&buf, "%d: %d\n", level, p.levelMaxBytes[level])
   142  				}
   143  				return buf.String()
   144  
   145  			default:
   146  				return fmt.Sprintf("unknown command: %s", d.Cmd)
   147  			}
   148  		})
   149  }
   150  
   151  func TestCompactionPickerTargetLevel(t *testing.T) {
   152  	var vers *version
   153  	var opts *Options
   154  	var pickerByScore *compactionPickerByScore
   155  
   156  	parseInProgress := func(vals []string) ([]compactionInfo, error) {
   157  		var levels []int
   158  		for _, s := range vals {
   159  			l, err := strconv.ParseInt(s, 10, 8)
   160  			if err != nil {
   161  				return nil, err
   162  			}
   163  			levels = append(levels, int(l))
   164  		}
   165  		if len(levels)%2 != 0 {
   166  			return nil, errors.New("odd number of levels with ongoing compactions")
   167  		}
   168  		var inProgress []compactionInfo
   169  		for i := 0; i < len(levels); i += 2 {
   170  			inProgress = append(inProgress, compactionInfo{
   171  				inputs: []compactionLevel{
   172  					{level: levels[i]},
   173  					{level: levels[i+1]},
   174  				},
   175  				outputLevel: levels[i+1],
   176  			})
   177  		}
   178  		return inProgress, nil
   179  	}
   180  
   181  	resetCompacting := func() {
   182  		for _, files := range vers.Levels {
   183  			files.Slice().Each(func(f *fileMetadata) {
   184  				f.CompactionState = manifest.CompactionStateNotCompacting
   185  			})
   186  		}
   187  	}
   188  
   189  	datadriven.RunTest(t, "testdata/compaction_picker_target_level",
   190  		func(t *testing.T, d *datadriven.TestData) string {
   191  			switch d.Cmd {
   192  			case "init":
   193  				// loadVersion expects a single datadriven argument that it
   194  				// sets as Options.LBaseMaxBytes. It parses the input as
   195  				// newline-separated levels, specifying the level's file size
   196  				// and optionally additional compensation to be added during
   197  				// compensated file size calculations. Eg:
   198  				//
   199  				// init <LBaseMaxBytes>
   200  				// <level>: <size> [compensation]
   201  				// <level>: <size> [compensation]
   202  				var errMsg string
   203  				vers, opts, errMsg = loadVersion(t, d)
   204  				if errMsg != "" {
   205  					return errMsg
   206  				}
   207  				return runVersionFileSizes(vers)
   208  			case "init_cp":
   209  				resetCompacting()
   210  
   211  				var inProgress []compactionInfo
   212  				if arg, ok := d.Arg("ongoing"); ok {
   213  					var err error
   214  					inProgress, err = parseInProgress(arg.Vals)
   215  					if err != nil {
   216  						return err.Error()
   217  					}
   218  				}
   219  
   220  				p := newCompactionPicker(vers, opts, inProgress)
   221  				var ok bool
   222  				pickerByScore, ok = p.(*compactionPickerByScore)
   223  				require.True(t, ok)
   224  				return fmt.Sprintf("base: %d", pickerByScore.baseLevel)
   225  			case "queue":
   226  				var b strings.Builder
   227  				var inProgress []compactionInfo
   228  				for {
   229  					env := compactionEnv{
   230  						diskAvailBytes:          math.MaxUint64,
   231  						earliestUnflushedSeqNum: InternalKeySeqNumMax,
   232  						inProgressCompactions:   inProgress,
   233  					}
   234  					pc := pickerByScore.pickAuto(env)
   235  					if pc == nil {
   236  						break
   237  					}
   238  					fmt.Fprintf(&b, "L%d->L%d: %.1f\n", pc.startLevel.level, pc.outputLevel.level, pc.score)
   239  					inProgress = append(inProgress, compactionInfo{
   240  						inputs:      pc.inputs,
   241  						outputLevel: pc.outputLevel.level,
   242  						smallest:    pc.smallest,
   243  						largest:     pc.largest,
   244  					})
   245  					if pc.outputLevel.level == 0 {
   246  						// Once we pick one L0->L0 compaction, we'll keep on doing so
   247  						// because the test isn't marking files as Compacting.
   248  						break
   249  					}
   250  					for _, cl := range pc.inputs {
   251  						cl.files.Each(func(f *fileMetadata) {
   252  							f.CompactionState = manifest.CompactionStateCompacting
   253  							fmt.Fprintf(&b, "  %s marked as compacting\n", f)
   254  						})
   255  					}
   256  				}
   257  
   258  				resetCompacting()
   259  				return b.String()
   260  			case "pick":
   261  				resetCompacting()
   262  
   263  				var inProgress []compactionInfo
   264  				if len(d.CmdArgs) == 1 {
   265  					arg := d.CmdArgs[0]
   266  					if arg.Key != "ongoing" {
   267  						return "unknown arg: " + arg.Key
   268  					}
   269  					var err error
   270  					inProgress, err = parseInProgress(arg.Vals)
   271  					if err != nil {
   272  						return err.Error()
   273  					}
   274  				}
   275  
   276  				// Mark files as compacting for each in-progress compaction.
   277  				for i := range inProgress {
   278  					c := &inProgress[i]
   279  					for j, cl := range c.inputs {
   280  						iter := vers.Levels[cl.level].Iter()
   281  						for f := iter.First(); f != nil; f = iter.Next() {
   282  							if !f.IsCompacting() {
   283  								f.CompactionState = manifest.CompactionStateCompacting
   284  								c.inputs[j].files = iter.Take().Slice()
   285  								break
   286  							}
   287  						}
   288  					}
   289  					if c.inputs[0].level == 0 && c.outputLevel != 0 {
   290  						// L0->Lbase: mark all of Lbase as compacting.
   291  						c.inputs[1].files = vers.Levels[c.outputLevel].Slice()
   292  						for _, in := range c.inputs {
   293  							in.files.Each(func(f *fileMetadata) {
   294  								f.CompactionState = manifest.CompactionStateCompacting
   295  							})
   296  						}
   297  					}
   298  				}
   299  
   300  				var b strings.Builder
   301  				fmt.Fprintf(&b, "Initial state before pick:\n%s", runVersionFileSizes(vers))
   302  				pc := pickerByScore.pickAuto(compactionEnv{
   303  					earliestUnflushedSeqNum: InternalKeySeqNumMax,
   304  					inProgressCompactions:   inProgress,
   305  				})
   306  				if pc != nil {
   307  					fmt.Fprintf(&b, "Picked: L%d->L%d: %0.1f\n", pc.startLevel.level, pc.outputLevel.level, pc.score)
   308  				}
   309  				if pc == nil {
   310  					fmt.Fprintln(&b, "Picked: no compaction")
   311  				}
   312  				return b.String()
   313  			case "pick_manual":
   314  				var startLevel int
   315  				var start, end string
   316  				d.MaybeScanArgs(t, "level", &startLevel)
   317  				d.MaybeScanArgs(t, "start", &start)
   318  				d.MaybeScanArgs(t, "end", &end)
   319  
   320  				iStart := base.MakeInternalKey([]byte(start), InternalKeySeqNumMax, InternalKeyKindMax)
   321  				iEnd := base.MakeInternalKey([]byte(end), 0, 0)
   322  				manual := &manualCompaction{
   323  					done:  make(chan error, 1),
   324  					level: startLevel,
   325  					start: iStart.UserKey,
   326  					end:   iEnd.UserKey,
   327  				}
   328  
   329  				pc, retryLater := pickManualCompaction(
   330  					pickerByScore.vers,
   331  					opts,
   332  					compactionEnv{
   333  						earliestUnflushedSeqNum: InternalKeySeqNumMax,
   334  					},
   335  					pickerByScore.getBaseLevel(),
   336  					manual)
   337  				if pc == nil {
   338  					return fmt.Sprintf("nil, retryLater = %v", retryLater)
   339  				}
   340  
   341  				return fmt.Sprintf("L%d->L%d, retryLater = %v", pc.startLevel.level, pc.outputLevel.level, retryLater)
   342  			default:
   343  				return fmt.Sprintf("unknown command: %s", d.Cmd)
   344  			}
   345  		})
   346  }
   347  
   348  func TestCompactionPickerEstimatedCompactionDebt(t *testing.T) {
   349  	datadriven.RunTest(t, "testdata/compaction_picker_estimated_debt",
   350  		func(t *testing.T, d *datadriven.TestData) string {
   351  			switch d.Cmd {
   352  			case "init":
   353  				vers, opts, errMsg := loadVersion(t, d)
   354  				if errMsg != "" {
   355  					return errMsg
   356  				}
   357  				opts.MemTableSize = 1000
   358  
   359  				p := newCompactionPicker(vers, opts, nil)
   360  				return fmt.Sprintf("%d\n", p.estimatedCompactionDebt(0))
   361  
   362  			default:
   363  				return fmt.Sprintf("unknown command: %s", d.Cmd)
   364  			}
   365  		})
   366  }
   367  
   368  func TestCompactionPickerL0(t *testing.T) {
   369  	opts := (*Options)(nil).EnsureDefaults()
   370  	opts.Experimental.L0CompactionConcurrency = 1
   371  
   372  	parseMeta := func(s string) (*fileMetadata, error) {
   373  		parts := strings.Split(s, ":")
   374  		fileNum, err := strconv.Atoi(parts[0])
   375  		if err != nil {
   376  			return nil, err
   377  		}
   378  		fields := strings.Fields(parts[1])
   379  		parts = strings.Split(fields[0], "-")
   380  		if len(parts) != 2 {
   381  			return nil, errors.Errorf("malformed table spec: %s", s)
   382  		}
   383  		m := (&fileMetadata{
   384  			FileNum: base.FileNum(fileNum),
   385  		}).ExtendPointKeyBounds(
   386  			opts.Comparer.Compare,
   387  			base.ParseInternalKey(strings.TrimSpace(parts[0])),
   388  			base.ParseInternalKey(strings.TrimSpace(parts[1])),
   389  		)
   390  		m.SmallestSeqNum = m.Smallest.SeqNum()
   391  		m.LargestSeqNum = m.Largest.SeqNum()
   392  		m.InitPhysicalBacking()
   393  		return m, nil
   394  	}
   395  
   396  	var picker *compactionPickerByScore
   397  	var inProgressCompactions []compactionInfo
   398  	var pc *pickedCompaction
   399  
   400  	datadriven.RunTest(t, "testdata/compaction_picker_L0", func(t *testing.T, td *datadriven.TestData) string {
   401  		switch td.Cmd {
   402  		case "define":
   403  			fileMetas := [manifest.NumLevels][]*fileMetadata{}
   404  			baseLevel := manifest.NumLevels - 1
   405  			level := 0
   406  			var err error
   407  			lines := strings.Split(td.Input, "\n")
   408  			var compactionLines []string
   409  
   410  			for len(lines) > 0 {
   411  				data := strings.TrimSpace(lines[0])
   412  				lines = lines[1:]
   413  				switch data {
   414  				case "L0", "L1", "L2", "L3", "L4", "L5", "L6":
   415  					level, err = strconv.Atoi(data[1:])
   416  					if err != nil {
   417  						return err.Error()
   418  					}
   419  				case "compactions":
   420  					compactionLines, lines = lines, nil
   421  				default:
   422  					meta, err := parseMeta(data)
   423  					if err != nil {
   424  						return err.Error()
   425  					}
   426  					if level != 0 && level < baseLevel {
   427  						baseLevel = level
   428  					}
   429  					fileMetas[level] = append(fileMetas[level], meta)
   430  				}
   431  			}
   432  
   433  			// Parse in-progress compactions in the form of:
   434  			//   L0 000001 -> L2 000005
   435  			inProgressCompactions = nil
   436  			for len(compactionLines) > 0 {
   437  				parts := strings.Fields(compactionLines[0])
   438  				compactionLines = compactionLines[1:]
   439  
   440  				var level int
   441  				var info compactionInfo
   442  				first := true
   443  				compactionFiles := map[int][]*fileMetadata{}
   444  				for _, p := range parts {
   445  					switch p {
   446  					case "L0", "L1", "L2", "L3", "L4", "L5", "L6":
   447  						var err error
   448  						level, err = strconv.Atoi(p[1:])
   449  						if err != nil {
   450  							return err.Error()
   451  						}
   452  						if len(info.inputs) > 0 && info.inputs[len(info.inputs)-1].level == level {
   453  							// eg, L0 -> L0 compaction or L6 -> L6 compaction
   454  							continue
   455  						}
   456  						if info.outputLevel < level {
   457  							info.outputLevel = level
   458  						}
   459  						info.inputs = append(info.inputs, compactionLevel{level: level})
   460  					case "->":
   461  						continue
   462  					default:
   463  						fileNum, err := strconv.Atoi(p)
   464  						if err != nil {
   465  							return err.Error()
   466  						}
   467  						var compactFile *fileMetadata
   468  						for _, m := range fileMetas[level] {
   469  							if m.FileNum == FileNum(fileNum) {
   470  								compactFile = m
   471  							}
   472  						}
   473  						if compactFile == nil {
   474  							return fmt.Sprintf("cannot find compaction file %s", FileNum(fileNum))
   475  						}
   476  						compactFile.CompactionState = manifest.CompactionStateCompacting
   477  						if first || base.InternalCompare(DefaultComparer.Compare, info.largest, compactFile.Largest) < 0 {
   478  							info.largest = compactFile.Largest
   479  						}
   480  						if first || base.InternalCompare(DefaultComparer.Compare, info.smallest, compactFile.Smallest) > 0 {
   481  							info.smallest = compactFile.Smallest
   482  						}
   483  						first = false
   484  						compactionFiles[level] = append(compactionFiles[level], compactFile)
   485  					}
   486  				}
   487  				for i, cl := range info.inputs {
   488  					files := compactionFiles[cl.level]
   489  					info.inputs[i].files = manifest.NewLevelSliceSeqSorted(files)
   490  					// Mark as intra-L0 compacting if the compaction is
   491  					// L0 -> L0.
   492  					if info.outputLevel == 0 {
   493  						for _, f := range files {
   494  							f.IsIntraL0Compacting = true
   495  						}
   496  					}
   497  				}
   498  				inProgressCompactions = append(inProgressCompactions, info)
   499  			}
   500  
   501  			version := newVersion(opts, fileMetas)
   502  			version.L0Sublevels.InitCompactingFileInfo(inProgressL0Compactions(inProgressCompactions))
   503  			vs := &versionSet{
   504  				opts:    opts,
   505  				cmp:     DefaultComparer.Compare,
   506  				cmpName: DefaultComparer.Name,
   507  			}
   508  			vs.versions.Init(nil)
   509  			vs.append(version)
   510  			picker = &compactionPickerByScore{
   511  				opts:      opts,
   512  				vers:      version,
   513  				baseLevel: baseLevel,
   514  			}
   515  			vs.picker = picker
   516  			picker.initLevelMaxBytes(inProgressCompactions)
   517  
   518  			var buf bytes.Buffer
   519  			fmt.Fprint(&buf, version.String())
   520  			if len(inProgressCompactions) > 0 {
   521  				fmt.Fprintln(&buf, "compactions")
   522  				for _, c := range inProgressCompactions {
   523  					fmt.Fprintf(&buf, "  %s\n", c.String())
   524  				}
   525  			}
   526  			return buf.String()
   527  		case "pick-auto":
   528  			td.MaybeScanArgs(t, "l0_compaction_threshold", &opts.L0CompactionThreshold)
   529  			td.MaybeScanArgs(t, "l0_compaction_file_threshold", &opts.L0CompactionFileThreshold)
   530  
   531  			pc = picker.pickAuto(compactionEnv{
   532  				diskAvailBytes:          math.MaxUint64,
   533  				earliestUnflushedSeqNum: math.MaxUint64,
   534  				inProgressCompactions:   inProgressCompactions,
   535  			})
   536  			var result strings.Builder
   537  			if pc != nil {
   538  				checkClone(t, pc)
   539  				c := newCompaction(pc, opts, time.Now(), nil /* provider */)
   540  				fmt.Fprintf(&result, "L%d -> L%d\n", pc.startLevel.level, pc.outputLevel.level)
   541  				fmt.Fprintf(&result, "L%d: %s\n", pc.startLevel.level, fileNums(pc.startLevel.files))
   542  				if !pc.outputLevel.files.Empty() {
   543  					fmt.Fprintf(&result, "L%d: %s\n", pc.outputLevel.level, fileNums(pc.outputLevel.files))
   544  				}
   545  				if !c.grandparents.Empty() {
   546  					fmt.Fprintf(&result, "grandparents: %s\n", fileNums(c.grandparents))
   547  				}
   548  			} else {
   549  				return "nil"
   550  			}
   551  			return result.String()
   552  		case "mark-for-compaction":
   553  			var fileNum uint64
   554  			td.ScanArgs(t, "file", &fileNum)
   555  			for l, lm := range picker.vers.Levels {
   556  				iter := lm.Iter()
   557  				for f := iter.First(); f != nil; f = iter.Next() {
   558  					if f.FileNum != base.FileNum(fileNum) {
   559  						continue
   560  					}
   561  					f.MarkedForCompaction = true
   562  					picker.vers.Stats.MarkedForCompaction++
   563  					picker.vers.Levels[l].InvalidateAnnotation(markedForCompactionAnnotator{})
   564  					return fmt.Sprintf("marked L%d.%s", l, f.FileNum)
   565  				}
   566  			}
   567  			return "not-found"
   568  		case "max-output-file-size":
   569  			if pc == nil {
   570  				return "no compaction"
   571  			}
   572  			return fmt.Sprintf("%d", pc.maxOutputFileSize)
   573  		case "max-overlap-bytes":
   574  			if pc == nil {
   575  				return "no compaction"
   576  			}
   577  			return fmt.Sprintf("%d", pc.maxOverlapBytes)
   578  		}
   579  		return fmt.Sprintf("unrecognized command: %s", td.Cmd)
   580  	})
   581  }
   582  
   583  func TestCompactionPickerConcurrency(t *testing.T) {
   584  	opts := (*Options)(nil).EnsureDefaults()
   585  	opts.Experimental.L0CompactionConcurrency = 1
   586  
   587  	parseMeta := func(s string) (*fileMetadata, error) {
   588  		parts := strings.Split(s, ":")
   589  		fileNum, err := strconv.Atoi(parts[0])
   590  		if err != nil {
   591  			return nil, err
   592  		}
   593  		fields := strings.Fields(parts[1])
   594  		parts = strings.Split(fields[0], "-")
   595  		if len(parts) != 2 {
   596  			return nil, errors.Errorf("malformed table spec: %s", s)
   597  		}
   598  		m := (&fileMetadata{
   599  			FileNum: base.FileNum(fileNum),
   600  			Size:    1028,
   601  		}).ExtendPointKeyBounds(
   602  			opts.Comparer.Compare,
   603  			base.ParseInternalKey(strings.TrimSpace(parts[0])),
   604  			base.ParseInternalKey(strings.TrimSpace(parts[1])),
   605  		)
   606  		m.InitPhysicalBacking()
   607  		for _, p := range fields[1:] {
   608  			if strings.HasPrefix(p, "size=") {
   609  				v, err := strconv.Atoi(strings.TrimPrefix(p, "size="))
   610  				if err != nil {
   611  					return nil, err
   612  				}
   613  				m.Size = uint64(v)
   614  			}
   615  		}
   616  		m.SmallestSeqNum = m.Smallest.SeqNum()
   617  		m.LargestSeqNum = m.Largest.SeqNum()
   618  		return m, nil
   619  	}
   620  
   621  	var picker *compactionPickerByScore
   622  	var inProgressCompactions []compactionInfo
   623  
   624  	datadriven.RunTest(t, "testdata/compaction_picker_concurrency", func(t *testing.T, td *datadriven.TestData) string {
   625  		switch td.Cmd {
   626  		case "define":
   627  			fileMetas := [manifest.NumLevels][]*fileMetadata{}
   628  			level := 0
   629  			var err error
   630  			lines := strings.Split(td.Input, "\n")
   631  			var compactionLines []string
   632  
   633  			for len(lines) > 0 {
   634  				data := strings.TrimSpace(lines[0])
   635  				lines = lines[1:]
   636  				switch data {
   637  				case "L0", "L1", "L2", "L3", "L4", "L5", "L6":
   638  					level, err = strconv.Atoi(data[1:])
   639  					if err != nil {
   640  						return err.Error()
   641  					}
   642  				case "compactions":
   643  					compactionLines, lines = lines, nil
   644  				default:
   645  					meta, err := parseMeta(data)
   646  					if err != nil {
   647  						return err.Error()
   648  					}
   649  					fileMetas[level] = append(fileMetas[level], meta)
   650  				}
   651  			}
   652  
   653  			// Parse in-progress compactions in the form of:
   654  			//   L0 000001 -> L2 000005
   655  			inProgressCompactions = nil
   656  			for len(compactionLines) > 0 {
   657  				parts := strings.Fields(compactionLines[0])
   658  				compactionLines = compactionLines[1:]
   659  
   660  				var level int
   661  				var info compactionInfo
   662  				first := true
   663  				compactionFiles := map[int][]*fileMetadata{}
   664  				for _, p := range parts {
   665  					switch p {
   666  					case "L0", "L1", "L2", "L3", "L4", "L5", "L6":
   667  						var err error
   668  						level, err = strconv.Atoi(p[1:])
   669  						if err != nil {
   670  							return err.Error()
   671  						}
   672  						if len(info.inputs) > 0 && info.inputs[len(info.inputs)-1].level == level {
   673  							// eg, L0 -> L0 compaction or L6 -> L6 compaction
   674  							continue
   675  						}
   676  						if info.outputLevel < level {
   677  							info.outputLevel = level
   678  						}
   679  						info.inputs = append(info.inputs, compactionLevel{level: level})
   680  					case "->":
   681  						continue
   682  					default:
   683  						fileNum, err := strconv.Atoi(p)
   684  						if err != nil {
   685  							return err.Error()
   686  						}
   687  						var compactFile *fileMetadata
   688  						for _, m := range fileMetas[level] {
   689  							if m.FileNum == FileNum(fileNum) {
   690  								compactFile = m
   691  							}
   692  						}
   693  						if compactFile == nil {
   694  							return fmt.Sprintf("cannot find compaction file %s", FileNum(fileNum))
   695  						}
   696  						compactFile.CompactionState = manifest.CompactionStateCompacting
   697  						if first || base.InternalCompare(DefaultComparer.Compare, info.largest, compactFile.Largest) < 0 {
   698  							info.largest = compactFile.Largest
   699  						}
   700  						if first || base.InternalCompare(DefaultComparer.Compare, info.smallest, compactFile.Smallest) > 0 {
   701  							info.smallest = compactFile.Smallest
   702  						}
   703  						first = false
   704  						compactionFiles[level] = append(compactionFiles[level], compactFile)
   705  					}
   706  				}
   707  				for i, cl := range info.inputs {
   708  					files := compactionFiles[cl.level]
   709  					if cl.level == 0 {
   710  						info.inputs[i].files = manifest.NewLevelSliceSeqSorted(files)
   711  					} else {
   712  						info.inputs[i].files = manifest.NewLevelSliceKeySorted(DefaultComparer.Compare, files)
   713  					}
   714  					// Mark as intra-L0 compacting if the compaction is
   715  					// L0 -> L0.
   716  					if info.outputLevel == 0 {
   717  						for _, f := range files {
   718  							f.IsIntraL0Compacting = true
   719  						}
   720  					}
   721  				}
   722  				inProgressCompactions = append(inProgressCompactions, info)
   723  			}
   724  
   725  			version := newVersion(opts, fileMetas)
   726  			version.L0Sublevels.InitCompactingFileInfo(inProgressL0Compactions(inProgressCompactions))
   727  			vs := &versionSet{
   728  				opts:    opts,
   729  				cmp:     DefaultComparer.Compare,
   730  				cmpName: DefaultComparer.Name,
   731  			}
   732  			vs.versions.Init(nil)
   733  			vs.append(version)
   734  
   735  			picker = newCompactionPicker(version, opts, inProgressCompactions).(*compactionPickerByScore)
   736  			vs.picker = picker
   737  
   738  			var buf bytes.Buffer
   739  			fmt.Fprint(&buf, version.String())
   740  			if len(inProgressCompactions) > 0 {
   741  				fmt.Fprintln(&buf, "compactions")
   742  				for _, c := range inProgressCompactions {
   743  					fmt.Fprintf(&buf, "  %s\n", c.String())
   744  				}
   745  			}
   746  			return buf.String()
   747  
   748  		case "pick-auto":
   749  			td.MaybeScanArgs(t, "l0_compaction_threshold", &opts.L0CompactionThreshold)
   750  			td.MaybeScanArgs(t, "l0_compaction_concurrency", &opts.Experimental.L0CompactionConcurrency)
   751  			td.MaybeScanArgs(t, "compaction_debt_concurrency", &opts.Experimental.CompactionDebtConcurrency)
   752  
   753  			pc := picker.pickAuto(compactionEnv{
   754  				earliestUnflushedSeqNum: math.MaxUint64,
   755  				inProgressCompactions:   inProgressCompactions,
   756  			})
   757  			var result strings.Builder
   758  			if pc != nil {
   759  				c := newCompaction(pc, opts, time.Now(), nil /* provider */)
   760  				fmt.Fprintf(&result, "L%d -> L%d\n", pc.startLevel.level, pc.outputLevel.level)
   761  				fmt.Fprintf(&result, "L%d: %s\n", pc.startLevel.level, fileNums(pc.startLevel.files))
   762  				if !pc.outputLevel.files.Empty() {
   763  					fmt.Fprintf(&result, "L%d: %s\n", pc.outputLevel.level, fileNums(pc.outputLevel.files))
   764  				}
   765  				if !c.grandparents.Empty() {
   766  					fmt.Fprintf(&result, "grandparents: %s\n", fileNums(c.grandparents))
   767  				}
   768  			} else {
   769  				return "nil"
   770  			}
   771  			return result.String()
   772  		}
   773  		return fmt.Sprintf("unrecognized command: %s", td.Cmd)
   774  	})
   775  }
   776  
   777  func TestCompactionPickerPickReadTriggered(t *testing.T) {
   778  	opts := (*Options)(nil).EnsureDefaults()
   779  	var picker *compactionPickerByScore
   780  	var rcList readCompactionQueue
   781  	var vers *version
   782  
   783  	parseMeta := func(s string) (*fileMetadata, error) {
   784  		parts := strings.Split(s, ":")
   785  		fileNum, err := strconv.Atoi(parts[0])
   786  		if err != nil {
   787  			return nil, err
   788  		}
   789  		fields := strings.Fields(parts[1])
   790  		parts = strings.Split(fields[0], "-")
   791  		if len(parts) != 2 {
   792  			return nil, errors.Errorf("malformed table spec: %s. usage: <file-num>:start.SET.1-end.SET.2", s)
   793  		}
   794  		m := (&fileMetadata{
   795  			FileNum: base.FileNum(fileNum),
   796  			Size:    1028,
   797  		}).ExtendPointKeyBounds(
   798  			opts.Comparer.Compare,
   799  			base.ParseInternalKey(strings.TrimSpace(parts[0])),
   800  			base.ParseInternalKey(strings.TrimSpace(parts[1])),
   801  		)
   802  		m.InitPhysicalBacking()
   803  		for _, p := range fields[1:] {
   804  			if strings.HasPrefix(p, "size=") {
   805  				v, err := strconv.Atoi(strings.TrimPrefix(p, "size="))
   806  				if err != nil {
   807  					return nil, err
   808  				}
   809  				m.Size = uint64(v)
   810  			}
   811  		}
   812  		m.SmallestSeqNum = m.Smallest.SeqNum()
   813  		m.LargestSeqNum = m.Largest.SeqNum()
   814  		return m, nil
   815  	}
   816  
   817  	datadriven.RunTest(t, "testdata/compaction_picker_read_triggered", func(t *testing.T, td *datadriven.TestData) string {
   818  		switch td.Cmd {
   819  		case "define":
   820  			rcList = readCompactionQueue{}
   821  			fileMetas := [manifest.NumLevels][]*fileMetadata{}
   822  			level := 0
   823  			var err error
   824  			lines := strings.Split(td.Input, "\n")
   825  
   826  			for len(lines) > 0 {
   827  				data := strings.TrimSpace(lines[0])
   828  				lines = lines[1:]
   829  				switch data {
   830  				case "L0", "L1", "L2", "L3", "L4", "L5", "L6":
   831  					level, err = strconv.Atoi(data[1:])
   832  					if err != nil {
   833  						return err.Error()
   834  					}
   835  				default:
   836  					meta, err := parseMeta(data)
   837  					if err != nil {
   838  						return err.Error()
   839  					}
   840  					fileMetas[level] = append(fileMetas[level], meta)
   841  				}
   842  			}
   843  
   844  			vers = newVersion(opts, fileMetas)
   845  			vs := &versionSet{
   846  				opts:    opts,
   847  				cmp:     DefaultComparer.Compare,
   848  				cmpName: DefaultComparer.Name,
   849  			}
   850  			vs.versions.Init(nil)
   851  			vs.append(vers)
   852  			var inProgressCompactions []compactionInfo
   853  			picker = newCompactionPicker(vers, opts, inProgressCompactions).(*compactionPickerByScore)
   854  			vs.picker = picker
   855  
   856  			var buf bytes.Buffer
   857  			fmt.Fprint(&buf, vers.String())
   858  			return buf.String()
   859  
   860  		case "add-read-compaction":
   861  			for _, line := range strings.Split(td.Input, "\n") {
   862  				if line == "" {
   863  					continue
   864  				}
   865  				parts := strings.Split(line, " ")
   866  				if len(parts) != 3 {
   867  					return "error: malformed data for add-read-compaction. usage: <level>: <start>-<end> <filenum>"
   868  				}
   869  				if l, err := strconv.Atoi(parts[0][:1]); err == nil {
   870  					keys := strings.Split(parts[1], "-")
   871  					fileNum, _ := strconv.Atoi(parts[2])
   872  
   873  					rc := readCompaction{
   874  						level:   l,
   875  						start:   []byte(keys[0]),
   876  						end:     []byte(keys[1]),
   877  						fileNum: base.FileNum(fileNum),
   878  					}
   879  					rcList.add(&rc, DefaultComparer.Compare)
   880  				} else {
   881  					return err.Error()
   882  				}
   883  			}
   884  			return ""
   885  
   886  		case "show-read-compactions":
   887  			var sb strings.Builder
   888  			if rcList.size == 0 {
   889  				sb.WriteString("(none)")
   890  			}
   891  			for i := 0; i < rcList.size; i++ {
   892  				rc := rcList.at(i)
   893  				sb.WriteString(fmt.Sprintf("(level: %d, start: %s, end: %s)\n", rc.level, string(rc.start), string(rc.end)))
   894  			}
   895  			return sb.String()
   896  
   897  		case "pick-auto":
   898  			pc := picker.pickAuto(compactionEnv{
   899  				earliestUnflushedSeqNum: math.MaxUint64,
   900  				readCompactionEnv: readCompactionEnv{
   901  					readCompactions: &rcList,
   902  					flushing:        false,
   903  				},
   904  			})
   905  			var result strings.Builder
   906  			if pc != nil {
   907  				fmt.Fprintf(&result, "L%d -> L%d\n", pc.startLevel.level, pc.outputLevel.level)
   908  				fmt.Fprintf(&result, "L%d: %s\n", pc.startLevel.level, fileNums(pc.startLevel.files))
   909  				if !pc.outputLevel.files.Empty() {
   910  					fmt.Fprintf(&result, "L%d: %s\n", pc.outputLevel.level, fileNums(pc.outputLevel.files))
   911  				}
   912  			} else {
   913  				return "nil"
   914  			}
   915  			return result.String()
   916  		}
   917  		return fmt.Sprintf("unrecognized command: %s", td.Cmd)
   918  	})
   919  }
   920  
   921  type alwaysMultiLevel struct{}
   922  
   923  func (d alwaysMultiLevel) pick(
   924  	pcOrig *pickedCompaction, opts *Options, diskAvailBytes uint64,
   925  ) *pickedCompaction {
   926  	pcMulti := pcOrig.clone()
   927  	if !pcMulti.setupMultiLevelCandidate(opts, diskAvailBytes) {
   928  		return pcOrig
   929  	}
   930  	return pcMulti
   931  }
   932  
   933  func (d alwaysMultiLevel) allowL0() bool {
   934  	return false
   935  }
   936  
   937  func TestPickedCompactionSetupInputs(t *testing.T) {
   938  	opts := &Options{}
   939  	opts.EnsureDefaults()
   940  
   941  	parseMeta := func(s string) *fileMetadata {
   942  		parts := strings.Split(strings.TrimSpace(s), " ")
   943  		var fileSize uint64
   944  		var compacting bool
   945  		for _, part := range parts {
   946  			switch {
   947  			case part == "compacting":
   948  				compacting = true
   949  			case strings.HasPrefix(part, "size="):
   950  				v, err := strconv.ParseUint(strings.TrimPrefix(part, "size="), 10, 64)
   951  				require.NoError(t, err)
   952  				fileSize = v
   953  			}
   954  		}
   955  		tableParts := strings.Split(parts[0], "-")
   956  		if len(tableParts) != 2 {
   957  			t.Fatalf("malformed table spec: %s", s)
   958  		}
   959  		state := manifest.CompactionStateNotCompacting
   960  		if compacting {
   961  			state = manifest.CompactionStateCompacting
   962  		}
   963  		m := (&fileMetadata{
   964  			CompactionState: state,
   965  			Size:            fileSize,
   966  		}).ExtendPointKeyBounds(
   967  			opts.Comparer.Compare,
   968  			base.ParseInternalKey(strings.TrimSpace(tableParts[0])),
   969  			base.ParseInternalKey(strings.TrimSpace(tableParts[1])),
   970  		)
   971  		m.SmallestSeqNum = m.Smallest.SeqNum()
   972  		m.LargestSeqNum = m.Largest.SeqNum()
   973  		m.InitPhysicalBacking()
   974  		return m
   975  	}
   976  
   977  	setupInputTest := func(t *testing.T, d *datadriven.TestData) string {
   978  		switch d.Cmd {
   979  		case "setup-inputs":
   980  			var availBytes uint64 = math.MaxUint64
   981  			var maxLevelBytes [7]int64
   982  			args := d.CmdArgs
   983  
   984  			if len(args) > 0 && args[0].Key == "avail-bytes" {
   985  				require.Equal(t, 1, len(args[0].Vals))
   986  				var err error
   987  				availBytes, err = strconv.ParseUint(args[0].Vals[0], 10, 64)
   988  				require.NoError(t, err)
   989  				args = args[1:]
   990  			}
   991  
   992  			if len(args) != 2 {
   993  				return "setup-inputs [avail-bytes=XXX] <start> <end>"
   994  			}
   995  
   996  			pc := &pickedCompaction{
   997  				cmp:    DefaultComparer.Compare,
   998  				inputs: []compactionLevel{{level: -1}, {level: -1}},
   999  			}
  1000  			pc.startLevel, pc.outputLevel = &pc.inputs[0], &pc.inputs[1]
  1001  			var currentLevel int
  1002  			var files [numLevels][]*fileMetadata
  1003  			fileNum := FileNum(1)
  1004  
  1005  			for _, data := range strings.Split(d.Input, "\n") {
  1006  				switch data[:2] {
  1007  				case "L0", "L1", "L2", "L3", "L4", "L5", "L6":
  1008  					levelArgs := strings.Fields(data)
  1009  					level, err := strconv.Atoi(levelArgs[0][1:])
  1010  					if err != nil {
  1011  						return err.Error()
  1012  					}
  1013  					currentLevel = level
  1014  					if len(levelArgs) > 1 {
  1015  						maxSizeArg := strings.Replace(levelArgs[1], "max-size=", "", 1)
  1016  						maxSize, err := strconv.ParseInt(maxSizeArg, 10, 64)
  1017  						if err != nil {
  1018  							return err.Error()
  1019  						}
  1020  						maxLevelBytes[level] = maxSize
  1021  					} else {
  1022  						maxLevelBytes[level] = math.MaxInt64
  1023  					}
  1024  					if pc.startLevel.level == -1 {
  1025  						pc.startLevel.level = level
  1026  
  1027  					} else if pc.outputLevel.level == -1 {
  1028  						if pc.startLevel.level >= level {
  1029  							return fmt.Sprintf("startLevel=%d >= outputLevel=%d\n", pc.startLevel.level, level)
  1030  						}
  1031  						pc.outputLevel.level = level
  1032  					}
  1033  				default:
  1034  					meta := parseMeta(data)
  1035  					meta.FileNum = fileNum
  1036  					fileNum++
  1037  					files[currentLevel] = append(files[currentLevel], meta)
  1038  				}
  1039  			}
  1040  
  1041  			if pc.outputLevel.level == -1 {
  1042  				pc.outputLevel.level = pc.startLevel.level + 1
  1043  			}
  1044  			pc.version = newVersion(opts, files)
  1045  			pc.startLevel.files = pc.version.Overlaps(pc.startLevel.level, pc.cmp,
  1046  				[]byte(args[0].String()), []byte(args[1].String()), false /* exclusiveEnd */)
  1047  
  1048  			var isCompacting bool
  1049  			if !pc.setupInputs(opts, availBytes, pc.startLevel) {
  1050  				isCompacting = true
  1051  			}
  1052  			origPC := pc
  1053  			pc = pc.maybeAddLevel(opts, availBytes)
  1054  			// If pc points to a new pickedCompaction, a new multi level compaction
  1055  			// was initialized.
  1056  			initMultiLevel := pc != origPC
  1057  			checkClone(t, pc)
  1058  			var buf bytes.Buffer
  1059  			for _, cl := range pc.inputs {
  1060  				if cl.files.Empty() {
  1061  					continue
  1062  				}
  1063  
  1064  				fmt.Fprintf(&buf, "L%d\n", cl.level)
  1065  				cl.files.Each(func(f *fileMetadata) {
  1066  					fmt.Fprintf(&buf, "  %s\n", f)
  1067  				})
  1068  			}
  1069  			if isCompacting {
  1070  				fmt.Fprintf(&buf, "is-compacting\n")
  1071  			}
  1072  
  1073  			if initMultiLevel {
  1074  				extraLevel := pc.extraLevels[0].level
  1075  				fmt.Fprintf(&buf, "init-multi-level(%d,%d,%d)\n", pc.startLevel.level, extraLevel,
  1076  					pc.outputLevel.level)
  1077  				fmt.Fprintf(&buf, "Original WriteAmp %.2f; ML WriteAmp %.2f\n", origPC.predictedWriteAmp(), pc.predictedWriteAmp())
  1078  				fmt.Fprintf(&buf, "Original OverlappingRatio %.2f; ML OverlappingRatio %.2f\n", origPC.overlappingRatio(), pc.overlappingRatio())
  1079  			}
  1080  			return buf.String()
  1081  
  1082  		default:
  1083  			return fmt.Sprintf("unknown command: %s", d.Cmd)
  1084  		}
  1085  	}
  1086  
  1087  	t.Logf("Test basic setup inputs behavior without multi level compactions")
  1088  	opts.Experimental.MultiLevelCompactionHeuristic = NoMultiLevel{}
  1089  	datadriven.RunTest(t, "testdata/compaction_setup_inputs",
  1090  		setupInputTest)
  1091  
  1092  	t.Logf("Turning multi level compaction on")
  1093  	opts.Experimental.MultiLevelCompactionHeuristic = alwaysMultiLevel{}
  1094  	datadriven.RunTest(t, "testdata/compaction_setup_inputs_multilevel_dummy",
  1095  		setupInputTest)
  1096  
  1097  	t.Logf("Try Write-Amp Heuristic")
  1098  	opts.Experimental.MultiLevelCompactionHeuristic = WriteAmpHeuristic{}
  1099  	datadriven.RunTest(t, "testdata/compaction_setup_inputs_multilevel_write_amp",
  1100  		setupInputTest)
  1101  }
  1102  
  1103  func TestPickedCompactionExpandInputs(t *testing.T) {
  1104  	opts := &Options{}
  1105  	opts.EnsureDefaults()
  1106  	cmp := DefaultComparer.Compare
  1107  	var files []*fileMetadata
  1108  
  1109  	parseMeta := func(s string) *fileMetadata {
  1110  		parts := strings.Split(s, "-")
  1111  		if len(parts) != 2 {
  1112  			t.Fatalf("malformed table spec: %s", s)
  1113  		}
  1114  		m := (&fileMetadata{}).ExtendPointKeyBounds(
  1115  			opts.Comparer.Compare,
  1116  			base.ParseInternalKey(parts[0]),
  1117  			base.ParseInternalKey(parts[1]),
  1118  		)
  1119  		m.InitPhysicalBacking()
  1120  		return m
  1121  	}
  1122  
  1123  	datadriven.RunTest(t, "testdata/compaction_expand_inputs",
  1124  		func(t *testing.T, d *datadriven.TestData) string {
  1125  			switch d.Cmd {
  1126  			case "define":
  1127  				files = nil
  1128  				if len(d.Input) == 0 {
  1129  					return ""
  1130  				}
  1131  				for _, data := range strings.Split(d.Input, "\n") {
  1132  					meta := parseMeta(data)
  1133  					meta.FileNum = FileNum(len(files))
  1134  					files = append(files, meta)
  1135  				}
  1136  				manifest.SortBySmallest(files, cmp)
  1137  				return ""
  1138  
  1139  			case "expand-inputs":
  1140  				pc := &pickedCompaction{
  1141  					cmp:    cmp,
  1142  					inputs: []compactionLevel{{level: 1}},
  1143  				}
  1144  				pc.startLevel = &pc.inputs[0]
  1145  
  1146  				var filesLevelled [numLevels][]*fileMetadata
  1147  				filesLevelled[pc.startLevel.level] = files
  1148  				pc.version = newVersion(opts, filesLevelled)
  1149  
  1150  				if len(d.CmdArgs) != 1 {
  1151  					return fmt.Sprintf("%s expects 1 argument", d.Cmd)
  1152  				}
  1153  				index, err := strconv.ParseInt(d.CmdArgs[0].String(), 10, 64)
  1154  				if err != nil {
  1155  					return err.Error()
  1156  				}
  1157  
  1158  				// Advance the iterator to position `index`.
  1159  				iter := pc.version.Levels[pc.startLevel.level].Iter()
  1160  				_ = iter.First()
  1161  				for i := int64(0); i < index; i++ {
  1162  					_ = iter.Next()
  1163  				}
  1164  
  1165  				inputs, _ := expandToAtomicUnit(cmp, iter.Take().Slice(), true /* disableIsCompacting */)
  1166  
  1167  				var buf bytes.Buffer
  1168  				inputs.Each(func(f *fileMetadata) {
  1169  					fmt.Fprintf(&buf, "%d: %s-%s\n", f.FileNum, f.Smallest, f.Largest)
  1170  				})
  1171  				return buf.String()
  1172  
  1173  			default:
  1174  				return fmt.Sprintf("unknown command: %s", d.Cmd)
  1175  			}
  1176  		})
  1177  }
  1178  
  1179  func TestCompactionOutputFileSize(t *testing.T) {
  1180  	opts := (*Options)(nil).EnsureDefaults()
  1181  	var picker *compactionPickerByScore
  1182  	var vers *version
  1183  
  1184  	parseMeta := func(s string) (*fileMetadata, error) {
  1185  		parts := strings.Split(s, ":")
  1186  		fileNum, err := strconv.Atoi(parts[0])
  1187  		if err != nil {
  1188  			return nil, err
  1189  		}
  1190  		fields := strings.Fields(parts[1])
  1191  		parts = strings.Split(fields[0], "-")
  1192  		if len(parts) != 2 {
  1193  			return nil, errors.Errorf("malformed table spec: %s. usage: <file-num>:start.SET.1-end.SET.2", s)
  1194  		}
  1195  		m := (&fileMetadata{
  1196  			FileNum: base.FileNum(fileNum),
  1197  			Size:    1028,
  1198  		}).ExtendPointKeyBounds(
  1199  			opts.Comparer.Compare,
  1200  			base.ParseInternalKey(strings.TrimSpace(parts[0])),
  1201  			base.ParseInternalKey(strings.TrimSpace(parts[1])),
  1202  		)
  1203  		m.InitPhysicalBacking()
  1204  		for _, p := range fields[1:] {
  1205  			if strings.HasPrefix(p, "size=") {
  1206  				v, err := strconv.Atoi(strings.TrimPrefix(p, "size="))
  1207  				if err != nil {
  1208  					return nil, err
  1209  				}
  1210  				m.Size = uint64(v)
  1211  			}
  1212  			if strings.HasPrefix(p, "range-deletions-bytes-estimate=") {
  1213  				v, err := strconv.Atoi(strings.TrimPrefix(p, "range-deletions-bytes-estimate="))
  1214  				if err != nil {
  1215  					return nil, err
  1216  				}
  1217  				m.Stats.RangeDeletionsBytesEstimate = uint64(v)
  1218  				m.Stats.NumDeletions = 1 // At least one range del responsible for the deletion bytes.
  1219  				m.StatsMarkValid()
  1220  			}
  1221  		}
  1222  		m.SmallestSeqNum = m.Smallest.SeqNum()
  1223  		m.LargestSeqNum = m.Largest.SeqNum()
  1224  		return m, nil
  1225  	}
  1226  
  1227  	datadriven.RunTest(t, "testdata/compaction_output_file_size", func(t *testing.T, td *datadriven.TestData) string {
  1228  		switch td.Cmd {
  1229  		case "define":
  1230  			fileMetas := [manifest.NumLevels][]*fileMetadata{}
  1231  			level := 0
  1232  			var err error
  1233  			lines := strings.Split(td.Input, "\n")
  1234  
  1235  			for len(lines) > 0 {
  1236  				data := strings.TrimSpace(lines[0])
  1237  				lines = lines[1:]
  1238  				switch data {
  1239  				case "L0", "L1", "L2", "L3", "L4", "L5", "L6":
  1240  					level, err = strconv.Atoi(data[1:])
  1241  					if err != nil {
  1242  						return err.Error()
  1243  					}
  1244  				default:
  1245  					meta, err := parseMeta(data)
  1246  					if err != nil {
  1247  						return err.Error()
  1248  					}
  1249  					fileMetas[level] = append(fileMetas[level], meta)
  1250  				}
  1251  			}
  1252  
  1253  			vers = newVersion(opts, fileMetas)
  1254  			vs := &versionSet{
  1255  				opts:    opts,
  1256  				cmp:     DefaultComparer.Compare,
  1257  				cmpName: DefaultComparer.Name,
  1258  			}
  1259  			vs.versions.Init(nil)
  1260  			vs.append(vers)
  1261  			var inProgressCompactions []compactionInfo
  1262  			picker = newCompactionPicker(vers, opts, inProgressCompactions).(*compactionPickerByScore)
  1263  			vs.picker = picker
  1264  
  1265  			var buf bytes.Buffer
  1266  			fmt.Fprint(&buf, vers.String())
  1267  			return buf.String()
  1268  
  1269  		case "pick-auto":
  1270  			pc := picker.pickAuto(compactionEnv{
  1271  				earliestUnflushedSeqNum: math.MaxUint64,
  1272  				earliestSnapshotSeqNum:  math.MaxUint64,
  1273  			})
  1274  			var buf bytes.Buffer
  1275  			if pc != nil {
  1276  				fmt.Fprintf(&buf, "L%d -> L%d\n", pc.startLevel.level, pc.outputLevel.level)
  1277  				fmt.Fprintf(&buf, "L%d: %s\n", pc.startLevel.level, fileNums(pc.startLevel.files))
  1278  				fmt.Fprintf(&buf, "maxOutputFileSize: %d\n", pc.maxOutputFileSize)
  1279  			} else {
  1280  				return "nil"
  1281  			}
  1282  			return buf.String()
  1283  
  1284  		default:
  1285  			return fmt.Sprintf("unrecognized command: %s", td.Cmd)
  1286  		}
  1287  	})
  1288  }
  1289  
  1290  func TestCompactionPickerCompensatedSize(t *testing.T) {
  1291  	testCases := []struct {
  1292  		size                  uint64
  1293  		pointDelEstimateBytes uint64
  1294  		rangeDelEstimateBytes uint64
  1295  		wantBytes             uint64
  1296  	}{
  1297  		{
  1298  			size:                  100,
  1299  			pointDelEstimateBytes: 0,
  1300  			rangeDelEstimateBytes: 0,
  1301  			wantBytes:             100,
  1302  		},
  1303  		{
  1304  			size:                  100,
  1305  			pointDelEstimateBytes: 10,
  1306  			rangeDelEstimateBytes: 0,
  1307  			wantBytes:             100 + 10,
  1308  		},
  1309  		{
  1310  			size:                  100,
  1311  			pointDelEstimateBytes: 10,
  1312  			rangeDelEstimateBytes: 5,
  1313  			wantBytes:             100 + 10 + 5,
  1314  		},
  1315  	}
  1316  
  1317  	for _, tc := range testCases {
  1318  		t.Run("", func(t *testing.T) {
  1319  			f := &fileMetadata{Size: tc.size}
  1320  			f.InitPhysicalBacking()
  1321  			f.Stats.PointDeletionsBytesEstimate = tc.pointDelEstimateBytes
  1322  			f.Stats.RangeDeletionsBytesEstimate = tc.rangeDelEstimateBytes
  1323  			gotBytes := compensatedSize(f)
  1324  			require.Equal(t, tc.wantBytes, gotBytes)
  1325  		})
  1326  	}
  1327  }
  1328  
  1329  func TestCompactionPickerPickFile(t *testing.T) {
  1330  	fs := vfs.NewMem()
  1331  	opts := &Options{
  1332  		Comparer:           testkeys.Comparer,
  1333  		FormatMajorVersion: FormatNewest,
  1334  		FS:                 fs,
  1335  	}
  1336  
  1337  	d, err := Open("", opts)
  1338  	require.NoError(t, err)
  1339  	defer func() {
  1340  		if d != nil {
  1341  			require.NoError(t, d.Close())
  1342  		}
  1343  	}()
  1344  
  1345  	datadriven.RunTest(t, "testdata/compaction_picker_pick_file", func(t *testing.T, td *datadriven.TestData) string {
  1346  		switch td.Cmd {
  1347  		case "define":
  1348  			require.NoError(t, d.Close())
  1349  
  1350  			d, err = runDBDefineCmd(td, opts)
  1351  			if err != nil {
  1352  				return err.Error()
  1353  			}
  1354  			d.mu.Lock()
  1355  			s := d.mu.versions.currentVersion().String()
  1356  			d.mu.Unlock()
  1357  			return s
  1358  
  1359  		case "file-sizes":
  1360  			return runTableFileSizesCmd(td, d)
  1361  
  1362  		case "pick-file":
  1363  			s := strings.TrimPrefix(td.CmdArgs[0].String(), "L")
  1364  			level, err := strconv.Atoi(s)
  1365  			if err != nil {
  1366  				return fmt.Sprintf("unable to parse arg %q as level", td.CmdArgs[0].String())
  1367  			}
  1368  			if level == 0 {
  1369  				panic("L0 picking unimplemented")
  1370  			}
  1371  			d.mu.Lock()
  1372  			defer d.mu.Unlock()
  1373  
  1374  			// Use maybeScheduleCompactionPicker to take care of all of the
  1375  			// initialization of the compaction-picking environment, but never
  1376  			// pick a compaction; just call pickFile using the user-provided
  1377  			// level.
  1378  			var lf manifest.LevelFile
  1379  			var ok bool
  1380  			d.maybeScheduleCompactionPicker(func(untypedPicker compactionPicker, env compactionEnv) *pickedCompaction {
  1381  				p := untypedPicker.(*compactionPickerByScore)
  1382  				lf, ok = pickCompactionSeedFile(p.vers, opts, level, level+1, env.earliestSnapshotSeqNum)
  1383  				return nil
  1384  			})
  1385  			if !ok {
  1386  				return "(none)"
  1387  			}
  1388  			return lf.FileMetadata.String()
  1389  
  1390  		default:
  1391  			return fmt.Sprintf("unknown command: %s", td.Cmd)
  1392  		}
  1393  	})
  1394  }
  1395  
  1396  type pausableCleaner struct {
  1397  	mu      sync.Mutex
  1398  	cond    sync.Cond
  1399  	paused  bool
  1400  	cleaner Cleaner
  1401  }
  1402  
  1403  func (c *pausableCleaner) Clean(fs vfs.FS, fileType base.FileType, path string) error {
  1404  	c.mu.Lock()
  1405  	defer c.mu.Unlock()
  1406  	for c.paused {
  1407  		c.cond.Wait()
  1408  	}
  1409  	return c.cleaner.Clean(fs, fileType, path)
  1410  }
  1411  
  1412  func (c *pausableCleaner) pause() {
  1413  	c.mu.Lock()
  1414  	defer c.mu.Unlock()
  1415  	c.paused = true
  1416  }
  1417  
  1418  func (c *pausableCleaner) resume() {
  1419  	c.mu.Lock()
  1420  	defer c.mu.Unlock()
  1421  	c.paused = false
  1422  	c.cond.Broadcast()
  1423  }
  1424  
  1425  func TestCompactionPickerScores(t *testing.T) {
  1426  	fs := vfs.NewMem()
  1427  	cleaner := pausableCleaner{cleaner: DeleteCleaner{}}
  1428  	cleaner.cond.L = &cleaner.mu
  1429  	opts := &Options{
  1430  		Cleaner:                     &cleaner,
  1431  		Comparer:                    testkeys.Comparer,
  1432  		DisableAutomaticCompactions: true,
  1433  		FormatMajorVersion:          FormatNewest,
  1434  		FS:                          fs,
  1435  	}
  1436  
  1437  	d, err := Open("", opts)
  1438  	require.NoError(t, err)
  1439  	defer func() {
  1440  		if d != nil {
  1441  			cleaner.resume()
  1442  			require.NoError(t, closeAllSnapshots(d))
  1443  			require.NoError(t, d.Close())
  1444  		}
  1445  	}()
  1446  
  1447  	var buf bytes.Buffer
  1448  	datadriven.RunTest(t, "testdata/compaction_picker_scores", func(t *testing.T, td *datadriven.TestData) string {
  1449  		switch td.Cmd {
  1450  		case "define":
  1451  			require.NoError(t, closeAllSnapshots(d))
  1452  			require.NoError(t, d.Close())
  1453  
  1454  			if td.HasArg("pause-cleaning") {
  1455  				cleaner.pause()
  1456  			}
  1457  
  1458  			d, err = runDBDefineCmd(td, opts)
  1459  			if err != nil {
  1460  				return err.Error()
  1461  			}
  1462  			d.mu.Lock()
  1463  			s := d.mu.versions.currentVersion().String()
  1464  			d.mu.Unlock()
  1465  			return s
  1466  
  1467  		case "disable-table-stats":
  1468  			d.mu.Lock()
  1469  			d.opts.private.disableTableStats = true
  1470  			d.mu.Unlock()
  1471  			return ""
  1472  
  1473  		case "enable-table-stats":
  1474  			d.mu.Lock()
  1475  			d.opts.private.disableTableStats = false
  1476  			d.maybeCollectTableStatsLocked()
  1477  			d.mu.Unlock()
  1478  			return ""
  1479  
  1480  		case "resume-cleaning":
  1481  			cleaner.resume()
  1482  			return ""
  1483  
  1484  		case "ingest":
  1485  			if err = runBuildCmd(td, d, d.opts.FS); err != nil {
  1486  				return err.Error()
  1487  			}
  1488  			if err = runIngestCmd(td, d, d.opts.FS); err != nil {
  1489  				return err.Error()
  1490  			}
  1491  			d.mu.Lock()
  1492  			s := d.mu.versions.currentVersion().String()
  1493  			d.mu.Unlock()
  1494  			return s
  1495  
  1496  		case "lsm":
  1497  			return runLSMCmd(td, d)
  1498  
  1499  		case "maybe-compact":
  1500  			buf.Reset()
  1501  			d.mu.Lock()
  1502  			d.opts.DisableAutomaticCompactions = false
  1503  			d.maybeScheduleCompaction()
  1504  			fmt.Fprintf(&buf, "%d compactions in progress:", d.mu.compact.compactingCount)
  1505  			for c := range d.mu.compact.inProgress {
  1506  				fmt.Fprintf(&buf, "\n%s", c)
  1507  			}
  1508  			d.opts.DisableAutomaticCompactions = true
  1509  			d.mu.Unlock()
  1510  			return buf.String()
  1511  
  1512  		case "scores":
  1513  			waitFor := "completion"
  1514  			td.MaybeScanArgs(t, "wait-for-compaction", &waitFor)
  1515  
  1516  			// Wait for any running compactions to complete before calculating
  1517  			// scores. Otherwise, the output of this command is
  1518  			// nondeterministic.
  1519  			switch waitFor {
  1520  			case "completion":
  1521  				d.mu.Lock()
  1522  				for d.mu.compact.compactingCount > 0 {
  1523  					d.mu.compact.cond.Wait()
  1524  				}
  1525  				d.mu.Unlock()
  1526  			case "version-edit":
  1527  				func() {
  1528  					for {
  1529  						d.mu.Lock()
  1530  						wait := len(d.mu.compact.inProgress) > 0
  1531  						for c := range d.mu.compact.inProgress {
  1532  							wait = wait && !c.versionEditApplied
  1533  						}
  1534  						d.mu.Unlock()
  1535  						if !wait {
  1536  							return
  1537  						}
  1538  						// d.mu.compact.cond isn't notified until the compaction
  1539  						// is removed from inProgress, so we need to just sleep
  1540  						// and check again soon.
  1541  						time.Sleep(10 * time.Millisecond)
  1542  					}
  1543  				}()
  1544  			default:
  1545  				panic(fmt.Sprintf("unrecognized `wait-for-compaction` value: %q", waitFor))
  1546  			}
  1547  
  1548  			buf.Reset()
  1549  			fmt.Fprintf(&buf, "L       Size   Score\n")
  1550  			for l, lm := range d.Metrics().Levels {
  1551  				if l < numLevels-1 {
  1552  					fmt.Fprintf(&buf, "L%-3d\t%-7s%.1f\n", l, humanize.Bytes.Int64(lm.Size), lm.Score)
  1553  				} else {
  1554  					fmt.Fprintf(&buf, "L%-3d\t%-7s-\n", l, humanize.Bytes.Int64(lm.Size))
  1555  				}
  1556  			}
  1557  			return buf.String()
  1558  
  1559  		case "wait-pending-table-stats":
  1560  			return runTableStatsCmd(td, d)
  1561  
  1562  		default:
  1563  			return fmt.Sprintf("unknown command: %s", td.Cmd)
  1564  		}
  1565  	})
  1566  }
  1567  
  1568  func fileNums(files manifest.LevelSlice) string {
  1569  	var ss []string
  1570  	files.Each(func(f *fileMetadata) {
  1571  		ss = append(ss, f.FileNum.String())
  1572  	})
  1573  	sort.Strings(ss)
  1574  	return strings.Join(ss, ",")
  1575  }
  1576  
  1577  func checkClone(t *testing.T, pc *pickedCompaction) {
  1578  	pcClone := pc.clone()
  1579  	require.Equal(t, pc.String(), pcClone.String())
  1580  
  1581  	// ensure all input files are in new address
  1582  	for i := range pc.inputs {
  1583  		// Len could be zero if setup inputs rejected a level
  1584  		if pc.inputs[i].files.Len() > 0 {
  1585  			require.NotEqual(t, &pc.inputs[i], &pcClone.inputs[i])
  1586  		}
  1587  	}
  1588  	for i := range pc.startLevel.l0SublevelInfo {
  1589  		if pc.startLevel.l0SublevelInfo[i].Len() > 0 {
  1590  			require.NotEqual(t, &pc.startLevel.l0SublevelInfo[i], &pcClone.startLevel.l0SublevelInfo[i])
  1591  		}
  1592  	}
  1593  }