github.com/cockroachdb/pebble@v0.0.0-20231214172447-ab4952c5f87b/cmd/pebble/write_bench.go (about)

     1  // Copyright 2021 The LevelDB-Go and Pebble Authors. All rights reserved. Use
     2  // of this source code is governed by a BSD-style license that can be found in
     3  // the LICENSE file.
     4  
     5  package main
     6  
     7  import (
     8  	"context"
     9  	"fmt"
    10  	"sync"
    11  	"time"
    12  
    13  	"github.com/cockroachdb/pebble/internal/ackseq"
    14  	"github.com/cockroachdb/pebble/internal/randvar"
    15  	"github.com/cockroachdb/pebble/internal/rate"
    16  	"github.com/spf13/cobra"
    17  )
    18  
    19  // The following constants match the values that Cockroach uses in Admission
    20  // Control at the time of writing.
    21  // See: https://github.com/cockroachdb/cockroach/blob/cb5d5108a7705eac7be82bc7f0f8b6f4dc825b96/pkg/util/admission/granter.go#L1212-L1229
    22  const (
    23  	defaultL0FileLimit     = 1000
    24  	defaultL0SubLevelLimit = 20
    25  )
    26  
    27  var writeBenchConfig struct {
    28  	batch              *randvar.Flag
    29  	keys               string
    30  	values             *randvar.BytesFlag
    31  	concurrency        int
    32  	rateStart          int
    33  	incBase            int
    34  	testPeriod         time.Duration
    35  	cooloffPeriod      time.Duration
    36  	targetL0Files      int
    37  	targetL0SubLevels  int
    38  	maxRateDipFraction float64
    39  	debug              bool
    40  }
    41  
    42  var writeBenchCmd = &cobra.Command{
    43  	Use:   "write <dir>",
    44  	Short: "Run YCSB F to find an a sustainable write throughput",
    45  	Long: `
    46  Run YCSB F (100% writes) at varying levels of sustained write load (ops/sec) to
    47  determine an optimal value of write throughput.
    48  
    49  The benchmark works by maintaining a fixed amount of write load on the DB for a
    50  fixed amount of time. If the database can handle the sustained load - determined
    51  by a heuristic that takes into account the number of files in L0 sub-levels, the
    52  number of L0 sub-levels, and whether the DB has encountered a write stall (i.e.
    53  measured load on the DB drops to zero) - the load is increased on the DB.
    54  
    55  Load increases exponentially from an initial load. If the DB fails the heuristic
    56  at the given write load, the load on the DB is paused for a period of time (the
    57  cool-off period) before returning to the last value at which the DB could handle
    58  the load. The exponent is then reset and the process repeats from this new
    59  initial value. This allows the benchmark to converge on and oscillate around the
    60  optimal write load.
    61  
    62  The values of load at which the DB passes and fails the heuristic are maintained
    63  over the duration of the benchmark. On completion of the benchmark, an "optimal"
    64  value is computed. The optimal value is computed as the value that minimizes the
    65  mis-classification of the recorded "passes" and "fails"". This can be visualized
    66  as a point on the x-axis that separates the passes and fails into the left and
    67  right half-planes, minimizing the number of fails that fall to the left of this
    68  point (i.e. mis-classified fails) and the number of passes that fall to the
    69  right (i.e. mis-classified passes).
    70  
    71  The resultant "optimal sustained write load" value provides an estimate of the
    72  write load that the DB can sustain without failing the target heuristic.
    73  
    74  A typical invocation of the benchmark is as follows:
    75  
    76    pebble bench write [PATH] --wipe -c 1024 -d 8h --rate-start 30000 --debug
    77  `,
    78  	Args: cobra.ExactArgs(1),
    79  	RunE: runWriteBenchmark,
    80  }
    81  
    82  func init() {
    83  	initWriteBench(writeBenchCmd)
    84  }
    85  
    86  func initWriteBench(cmd *cobra.Command) {
    87  	// Default values for custom flags.
    88  	writeBenchConfig.batch = randvar.NewFlag("1")
    89  	writeBenchConfig.values = randvar.NewBytesFlag("1000")
    90  
    91  	cmd.Flags().Var(
    92  		writeBenchConfig.batch, "batch",
    93  		"batch size distribution [{zipf,uniform}:]min[-max]")
    94  	cmd.Flags().StringVar(
    95  		&writeBenchConfig.keys, "keys", "zipf", "latest, uniform, or zipf")
    96  	cmd.Flags().Var(
    97  		writeBenchConfig.values, "values",
    98  		"value size distribution [{zipf,uniform}:]min[-max][/<target-compression>]")
    99  	cmd.Flags().IntVarP(
   100  		&writeBenchConfig.concurrency, "concurrency", "c",
   101  		1, "number of concurrent workers")
   102  	cmd.Flags().IntVar(
   103  		&writeBenchConfig.rateStart, "rate-start",
   104  		1000, "starting write load (ops/sec)")
   105  	cmd.Flags().IntVar(
   106  		&writeBenchConfig.incBase, "rate-inc-base",
   107  		100, "increment / decrement base")
   108  	cmd.Flags().DurationVar(
   109  		&writeBenchConfig.testPeriod, "test-period",
   110  		60*time.Second, "time to run at a given write load")
   111  	cmd.Flags().DurationVar(
   112  		&writeBenchConfig.cooloffPeriod, "cooloff-period",
   113  		30*time.Second, "time to pause write load after a failure")
   114  	cmd.Flags().IntVar(
   115  		&writeBenchConfig.targetL0Files, "l0-files",
   116  		defaultL0FileLimit, "target L0 file count")
   117  	cmd.Flags().IntVar(
   118  		&writeBenchConfig.targetL0SubLevels, "l0-sublevels",
   119  		defaultL0SubLevelLimit, "target L0 sublevel count")
   120  	cmd.Flags().BoolVarP(
   121  		&wipe, "wipe", "w", false, "wipe the database before starting")
   122  	cmd.Flags().Float64Var(
   123  		&writeBenchConfig.maxRateDipFraction, "max-rate-dip-fraction", 0.1,
   124  		"fraction at which to mark a test-run as failed if the actual rate dips below (relative to the desired rate)")
   125  	cmd.Flags().BoolVar(
   126  		&writeBenchConfig.debug, "debug", false, "print benchmark debug information")
   127  }
   128  
   129  // writeBenchResult contains the results of a test run at a given rate. The
   130  // independent variable is the rate (in ops/sec) and the dependent variable is
   131  // whether the test passed or failed. Additional metadata associated with the
   132  // test run is also captured.
   133  type writeBenchResult struct {
   134  	name     string
   135  	rate     int           // The rate at which the test is currently running.
   136  	passed   bool          // Was the test successful at this rate.
   137  	elapsed  time.Duration // The total elapsed time of the test.
   138  	bytes    uint64        // The size of the LSM.
   139  	levels   int           // The number of levels occupied in the LSM.
   140  	writeAmp float64       // The write amplification.
   141  }
   142  
   143  // String implements fmt.Stringer, printing a raw benchmark line. These lines
   144  // are used when performing analysis on a given benchmark run.
   145  func (r writeBenchResult) String() string {
   146  	return fmt.Sprintf("BenchmarkRaw%s %d ops/sec %v pass %s elapsed %d bytes %d levels %.2f writeAmp",
   147  		r.name,
   148  		r.rate,
   149  		r.passed,
   150  		r.elapsed,
   151  		r.bytes,
   152  		r.levels,
   153  		r.writeAmp,
   154  	)
   155  }
   156  
   157  func runWriteBenchmark(_ *cobra.Command, args []string) error {
   158  	const workload = "F" // 100% inserts.
   159  	var (
   160  		writers      []*pauseWriter
   161  		writersWg    *sync.WaitGroup // Tracks completion of all pauseWriters.
   162  		cooloff      bool            // Is cool-off enabled.
   163  		streak       int             // The number of successive passes.
   164  		clockStart   time.Time       // Start time for current load.
   165  		cooloffStart time.Time       // When cool-off was enabled.
   166  		stack        []int           // Stack of passing load values.
   167  		pass, fail   []int           // Values of load that pass and fail, respectively.
   168  		rateAcc      float64         // Accumulator of measured rates for a single test run.
   169  	)
   170  
   171  	desiredRate := writeBenchConfig.rateStart
   172  	incBase := writeBenchConfig.incBase
   173  	weights, err := ycsbParseWorkload(workload)
   174  
   175  	if err != nil {
   176  		return err
   177  	}
   178  
   179  	keyDist, err := ycsbParseKeyDist(writeBenchConfig.keys)
   180  	if err != nil {
   181  		return err
   182  	}
   183  	batchDist := writeBenchConfig.batch
   184  	valueDist := writeBenchConfig.values
   185  
   186  	// Construct a new YCSB F benchmark with the configured values.
   187  	y := newYcsb(weights, keyDist, batchDist, nil /* scans */, valueDist)
   188  	y.keyNum = ackseq.New(0)
   189  
   190  	setLimit := func(l int) {
   191  		perWriterRate := float64(l) / float64(len(writers))
   192  		for _, w := range writers {
   193  			w.setRate(perWriterRate)
   194  		}
   195  	}
   196  
   197  	// Function closure to run on test-run failure.
   198  	onTestFail := func(r writeBenchResult, cancel func()) {
   199  		fail = append(fail, desiredRate)
   200  
   201  		// Emit a benchmark raw datapoint.
   202  		fmt.Println(r)
   203  
   204  		// We failed at the current load, we have two options:
   205  
   206  		// a) No room to backtrack. We're done.
   207  		if len(stack) == 0 {
   208  			debugPrint("no room to backtrack; exiting ...\n")
   209  			cancel()
   210  			writersWg.Wait()
   211  			return
   212  		}
   213  
   214  		// b) We still have room to backtrack. Reduce the load to the
   215  		// last known passing value.
   216  		desiredRate, stack = stack[len(stack)-1], stack[:len(stack)-1]
   217  		setLimit(desiredRate)
   218  
   219  		// Enter the cool-off period.
   220  		cooloff = true
   221  		var wg sync.WaitGroup
   222  		for _, w := range writers {
   223  			// With a large number of writers, pausing synchronously can
   224  			// take a material amount of time. Instead, pause the
   225  			// writers in parallel in the background, and wait for all
   226  			// to complete before continuing.
   227  			wg.Add(1)
   228  			go func(writer *pauseWriter) {
   229  				writer.pause()
   230  				wg.Done()
   231  			}(w)
   232  		}
   233  		wg.Wait()
   234  
   235  		// Reset the counters and clocks.
   236  		streak = 0
   237  		rateAcc = 0
   238  		cooloffStart = time.Now()
   239  		clockStart = time.Now()
   240  		debugPrint("Fail. Pausing writers for cool-off period.\n")
   241  		debugPrint(fmt.Sprintf("new rate=%d\npasses=%v\nfails=%v\nstack=%v\n",
   242  			desiredRate, pass, fail, stack))
   243  	}
   244  
   245  	// Function closure to run on test-run success.
   246  	onTestSuccess := func(r writeBenchResult) {
   247  		streak++
   248  		pass = append(pass, desiredRate)
   249  		stack = append(stack, desiredRate)
   250  
   251  		// Emit a benchmark raw datapoint.
   252  		r.passed = true
   253  		fmt.Println(r)
   254  
   255  		// Increase the rate.
   256  		desiredRate = desiredRate + incBase*(1<<(streak-1))
   257  		setLimit(desiredRate)
   258  
   259  		// Restart the test.
   260  		rateAcc = 0
   261  		clockStart = time.Now()
   262  
   263  		debugPrint(fmt.Sprintf("Pass.\nnew rate=%d\npasses=%v\nfails=%v\nstreak=%d\nstack=%v\n",
   264  			desiredRate, pass, fail, streak, stack))
   265  	}
   266  
   267  	name := fmt.Sprintf("write/values=%s", writeBenchConfig.values)
   268  	ctx, cancel := context.WithCancel(context.Background())
   269  	runTest(args[0], test{
   270  		init: func(db DB, wg *sync.WaitGroup) {
   271  			y.db = db
   272  			writersWg = wg
   273  
   274  			// Spawn the writers.
   275  			for i := 0; i < writeBenchConfig.concurrency; i++ {
   276  				writer := newPauseWriter(y, float64(desiredRate))
   277  				writers = append(writers, writer)
   278  				writersWg.Add(1)
   279  				go writer.run(ctx, wg)
   280  			}
   281  			setLimit(desiredRate)
   282  
   283  			// Start the clock on the current load.
   284  			clockStart = time.Now()
   285  		},
   286  		tick: func(elapsed time.Duration, i int) {
   287  			m := y.db.Metrics()
   288  			if i%20 == 0 {
   289  				if writeBenchConfig.debug && i > 0 {
   290  					fmt.Printf("%s\n", m)
   291  				}
   292  				fmt.Println("___elapsed___clock___rate(desired)___rate(actual)___L0files___L0levels___levels______lsmBytes___writeAmp")
   293  			}
   294  
   295  			// Print the current stats.
   296  			l0Files := m.Levels[0].NumFiles
   297  			l0Sublevels := m.Levels[0].Sublevels
   298  			nLevels := 0
   299  			for _, l := range m.Levels {
   300  				if l.BytesIn > 0 {
   301  					nLevels++
   302  				}
   303  			}
   304  			lsmBytes := m.DiskSpaceUsage()
   305  			total := m.Total()
   306  			writeAmp := (&total).WriteAmp()
   307  
   308  			var currRate float64
   309  			var stalled bool
   310  			y.reg.Tick(func(tick histogramTick) {
   311  				h := tick.Hist
   312  				currRate = float64(h.TotalCount()) / tick.Elapsed.Seconds()
   313  				stalled = !cooloff && currRate == 0
   314  			})
   315  			rateAcc += currRate
   316  
   317  			// The heuristic by which the DB can sustain a given write load is
   318  			// determined by whether the DB, for the configured window of time:
   319  			// 1) did not encounter a write stall (i.e. write load fell to
   320  			//    zero),
   321  			// 2) number of files in L0 was at or below the target, and
   322  			// 3) number of L0 sub-levels is at or below the target.
   323  			failed := stalled ||
   324  				int(l0Files) > writeBenchConfig.targetL0Files ||
   325  				int(l0Sublevels) > writeBenchConfig.targetL0SubLevels
   326  
   327  			// Print the result for this tick.
   328  			fmt.Printf("%10s %7s %15d %14.1f %9d %10d %8d %13d %10.2f\n",
   329  				time.Duration(elapsed.Seconds()+0.5)*time.Second,
   330  				time.Duration(time.Since(clockStart).Seconds()+0.5)*time.Second,
   331  				desiredRate,
   332  				currRate,
   333  				l0Files,
   334  				l0Sublevels,
   335  				nLevels,
   336  				lsmBytes,
   337  				writeAmp,
   338  			)
   339  
   340  			// If we're in cool-off mode, allow it to complete before resuming
   341  			// writing.
   342  			if cooloff {
   343  				if time.Since(cooloffStart) < writeBenchConfig.cooloffPeriod {
   344  					return
   345  				}
   346  				debugPrint("ending cool-off")
   347  
   348  				// Else, resume writing.
   349  				cooloff = false
   350  				for _, w := range writers {
   351  					w.unpause()
   352  				}
   353  				clockStart = time.Now()
   354  
   355  				return
   356  			}
   357  
   358  			r := writeBenchResult{
   359  				name:     name,
   360  				rate:     desiredRate,
   361  				elapsed:  time.Duration(elapsed.Seconds()+0.5) * time.Second,
   362  				bytes:    lsmBytes,
   363  				levels:   nLevels,
   364  				writeAmp: writeAmp,
   365  			}
   366  
   367  			if failed {
   368  				onTestFail(r, cancel)
   369  				return
   370  			}
   371  
   372  			// Else, the DB could handle the current load. We only increase
   373  			// after a fixed amount of time at this load as elapsed.
   374  			testElapsed := time.Since(clockStart)
   375  			if testElapsed < writeBenchConfig.testPeriod {
   376  				// This test-run still has time on the clock.
   377  				return
   378  			}
   379  
   380  			// This test-run has completed.
   381  
   382  			// If the average rate over the test is less than the desired rate,
   383  			// we mark this test-run as a failure. This handles cases where we
   384  			// encounter a bottleneck that limits write throughput but
   385  			// incorrectly mark the test as passed.
   386  			diff := 1 - rateAcc/(float64(desiredRate)*testElapsed.Seconds())
   387  			if diff > writeBenchConfig.maxRateDipFraction {
   388  				if writeBenchConfig.debug {
   389  					debugPrint(fmt.Sprintf(
   390  						"difference in rates (%.2f) exceeded threshold (%.2f); marking test as failed\n",
   391  						diff, writeBenchConfig.maxRateDipFraction,
   392  					))
   393  				}
   394  				onTestFail(r, cancel)
   395  				return
   396  			}
   397  
   398  			// Mark this test-run as passed.
   399  			onTestSuccess(r)
   400  		},
   401  		done: func(elapsed time.Duration) {
   402  			// Print final analysis.
   403  			var total int64
   404  			y.reg.Tick(func(tick histogramTick) {
   405  				total = tick.Cumulative.TotalCount()
   406  			})
   407  			fmt.Println("___elapsed___ops(total)")
   408  			fmt.Printf("%10s %12d\n", elapsed.Truncate(time.Second), total)
   409  		},
   410  	})
   411  
   412  	return nil
   413  }
   414  
   415  // debugPrint prints a debug line to stdout if debug logging is enabled via the
   416  // --debug flag.
   417  func debugPrint(s string) {
   418  	if !writeBenchConfig.debug {
   419  		return
   420  	}
   421  	fmt.Print("DEBUG: " + s)
   422  }
   423  
   424  // pauseWriter issues load against a pebble instance, and can be paused on
   425  // demand to allow the DB to recover.
   426  type pauseWriter struct {
   427  	y        *ycsb
   428  	limiter  *rate.Limiter
   429  	pauseC   chan struct{}
   430  	unpauseC chan struct{}
   431  }
   432  
   433  // newPauseWriter returns a new pauseWriter.
   434  func newPauseWriter(y *ycsb, initialRate float64) *pauseWriter {
   435  	// Set the burst rate for the limiter to the lowest sensible value to
   436  	// prevent excessive bursting. Note that a burst of zero effectively
   437  	// disables the rate limiter, as a wait time of +Inf is returned from all
   438  	// calls, and `wait(l *rate.Limiter)` will not sleep in this case.
   439  	const burst = 1
   440  	return &pauseWriter{
   441  		y:        y,
   442  		limiter:  rate.NewLimiter(float64(initialRate), burst),
   443  		pauseC:   make(chan struct{}),
   444  		unpauseC: make(chan struct{}),
   445  	}
   446  }
   447  
   448  // run starts the pauseWriter, issuing load against the DB.
   449  func (w *pauseWriter) run(ctx context.Context, wg *sync.WaitGroup) {
   450  	defer wg.Done()
   451  
   452  	buf := &ycsbBuf{rng: randvar.NewRand()}
   453  	hist := w.y.reg.Register("insert")
   454  	for {
   455  		select {
   456  		case <-ctx.Done():
   457  			return
   458  		case <-w.pauseC:
   459  			// Hold the goroutine here until we unpause.
   460  			<-w.unpauseC
   461  		default:
   462  			wait(w.limiter)
   463  			start := time.Now()
   464  			w.y.insert(w.y.db, buf)
   465  			hist.Record(time.Since(start))
   466  		}
   467  	}
   468  }
   469  
   470  // pause signals that the writer should pause after the current operation.
   471  func (w *pauseWriter) pause() {
   472  	w.pauseC <- struct{}{}
   473  }
   474  
   475  // unpause unpauses the writer.
   476  func (w *pauseWriter) unpause() {
   477  	w.unpauseC <- struct{}{}
   478  }
   479  
   480  // setRate sets the rate limit for this writer.
   481  func (w *pauseWriter) setRate(r float64) {
   482  	w.limiter.SetRate(r)
   483  }