github.com/cockroachdb/pebble@v1.1.2/tool/logs/compaction.go (about)

     1  // Copyright 2021 The LevelDB-Go and Pebble Authors. All rights reserved. Use
     2  // of this source code is governed by a BSD-style license that can be found in
     3  // the LICENSE file.
     4  
     5  package logs
     6  
     7  import (
     8  	"bufio"
     9  	"bytes"
    10  	"fmt"
    11  	"math"
    12  	"os"
    13  	"path/filepath"
    14  	"regexp"
    15  	"sort"
    16  	"strconv"
    17  	"strings"
    18  	"time"
    19  
    20  	"github.com/cockroachdb/errors"
    21  	"github.com/cockroachdb/pebble/internal/humanize"
    22  	"github.com/spf13/cobra"
    23  )
    24  
    25  var (
    26  	// Captures a common logging prefix that can be used as the context for the
    27  	// surrounding information captured by other expressions. Example:
    28  	//
    29  	//   I211215 14:26:56.012382 51831533 3@vendor/github.com/cockroachdb/pebble/compaction.go:1845 ⋮ [n5,pebble,s5] ...
    30  	//
    31  	logContextPattern = regexp.MustCompile(
    32  		`^.*` +
    33  			/* Timestamp        */ `(?P<timestamp>\d{6} \d{2}:\d{2}:\d{2}.\d{6}).*` +
    34  			/* Node / Store     */ `\[n(?P<node>\d+|\?).*,s(?P<store>\d+|\?).*?\].*`,
    35  	)
    36  	logContextPatternTimestampIdx = logContextPattern.SubexpIndex("timestamp")
    37  	logContextPatternNodeIdx      = logContextPattern.SubexpIndex("node")
    38  	logContextPatternStoreIdx     = logContextPattern.SubexpIndex("store")
    39  
    40  	// Matches either a compaction or a memtable flush log line.
    41  	//
    42  	// A compaction start / end line resembles:
    43  	//   "[JOB X] compact(ed|ing)"
    44  	//
    45  	// A memtable flush start / end line resembles:
    46  	//   "[JOB X] flush(ed|ing)"
    47  	//
    48  	// An ingested sstable flush looks like:
    49  	//   "[JOB 226] flushed 6 ingested flushables"
    50  	sentinelPattern          = regexp.MustCompile(`\[JOB.*(?P<prefix>compact|flush|ingest)(?P<suffix>ed|ing)[^:]`)
    51  	sentinelPatternPrefixIdx = sentinelPattern.SubexpIndex("prefix")
    52  	sentinelPatternSuffixIdx = sentinelPattern.SubexpIndex("suffix")
    53  
    54  	// Example compaction start and end log lines:
    55  	// 23.1 and older:
    56  	//   I211215 14:26:56.012382 51831533 3@vendor/github.com/cockroachdb/pebble/compaction.go:1845 ⋮ [n5,pebble,s5] 1216510  [JOB 284925] compacting(default) L2 [442555] (4.2 M) + L3 [445853] (8.4 M)
    57  	//   I211215 14:26:56.318543 51831533 3@vendor/github.com/cockroachdb/pebble/compaction.go:1886 ⋮ [n5,pebble,s5] 1216554  [JOB 284925] compacted(default) L2 [442555] (4.2 M) + L3 [445853] (8.4 M) -> L3 [445883 445887] (13 M), in 0.3s, output rate 42 M/s
    58  	// current:
    59  	//   I211215 14:26:56.012382 51831533 3@vendor/github.com/cockroachdb/pebble/compaction.go:1845 ⋮ [n5,pebble,s5] 1216510  [JOB 284925] compacting(default) L2 [442555] (4.2MB) + L3 [445853] (8.4MB)
    60  	//   I211215 14:26:56.318543 51831533 3@vendor/github.com/cockroachdb/pebble/compaction.go:1886 ⋮ [n5,pebble,s5] 1216554  [JOB 284925] compacted(default) L2 [442555] (4.2MB) + L3 [445853] (8.4MB) -> L3 [445883 445887] (13MB), in 0.3s, output rate 42MB/s
    61  	//
    62  	// NOTE: we use the log timestamp to compute the compaction duration rather
    63  	// than the Pebble log output.
    64  	compactionPattern = regexp.MustCompile(
    65  		`^.*` +
    66  			/* Job ID            */ `\[JOB (?P<job>\d+)]\s` +
    67  			/* Start / end       */ `compact(?P<suffix>ed|ing)` +
    68  
    69  			/* Compaction type   */
    70  			`\((?P<type>.*?)\)\s` +
    71  			/* Optional annotation*/ `?(\s*\[(?P<annotations>.*?)\]\s*)?` +
    72  
    73  			/* Start / end level */
    74  			`(?P<levels>L(?P<from>\d).*?(?:.*(?:\+|->)\sL(?P<to>\d))?` +
    75  			/* Bytes             */
    76  			`(?:.*?\((?P<bytes>[0-9.]+( [BKMGTPE]|[KMGTPE]?B))\))` +
    77  			/* Score */
    78  			`?(\s*(Score=\d+(\.\d+)))?)`,
    79  	)
    80  	compactionPatternJobIdx    = compactionPattern.SubexpIndex("job")
    81  	compactionPatternSuffixIdx = compactionPattern.SubexpIndex("suffix")
    82  	compactionPatternTypeIdx   = compactionPattern.SubexpIndex("type")
    83  	compactionPatternLevels    = compactionPattern.SubexpIndex("levels")
    84  	compactionPatternFromIdx   = compactionPattern.SubexpIndex("from")
    85  	compactionPatternToIdx     = compactionPattern.SubexpIndex("to")
    86  	compactionPatternBytesIdx  = compactionPattern.SubexpIndex("bytes")
    87  
    88  	// Example memtable flush log lines:
    89  	// 23.1 and older:
    90  	//   I211213 16:23:48.903751 21136 3@vendor/github.com/cockroachdb/pebble/event.go:599 ⋮ [n9,pebble,s9] 24 [JOB 10] flushing 2 memtables to L0
    91  	//   I211213 16:23:49.134464 21136 3@vendor/github.com/cockroachdb/pebble/event.go:603 ⋮ [n9,pebble,s9] 26 [JOB 10] flushed 2 memtables to L0 [1535806] (1.3 M), in 0.2s, output rate 5.8 M/s
    92  	// current:
    93  	//   I211213 16:23:48.903751 21136
    94  	//   3@vendor/github.com/cockroachdb/pebble/event.go:599 ⋮ [n9,pebble,s9] 24 [JOB 10] flushing 2 memtables (1.4MB)  to L0
    95  	//   I211213 16:23:49.134464 21136
    96  	//   3@vendor/github.com/cockroachdb/pebble/event.go:603 ⋮ [n9,pebble,s9] 26 [JOB 10] flushed 2 memtables (1.4MB) to L0 [1535806] (1.3MB), in 0.2s, output rate 5.8MB/s
    97  	//
    98  	// NOTE: we use the log timestamp to compute the flush duration rather than
    99  	// the Pebble log output.
   100  	flushPattern = regexp.MustCompile(
   101  		`^..*` +
   102  			/* Job ID                       */ `\[JOB (?P<job>\d+)]\s` +
   103  			/* Compaction type              */ `flush(?P<suffix>ed|ing)\s` +
   104  			/* Memtable count; size (23.2+) */ `\d+ memtables? (\([^)]+\))?` +
   105  			/* SSTable Bytes                */ `(?:.*?\((?P<bytes>[0-9.]+( [BKMGTPE]|[KMGTPE]?B))\))?`,
   106  	)
   107  	flushPatternSuffixIdx = flushPattern.SubexpIndex("suffix")
   108  	flushPatternJobIdx    = flushPattern.SubexpIndex("job")
   109  	flushPatternBytesIdx  = flushPattern.SubexpIndex("bytes")
   110  
   111  	// Example ingested log lines:
   112  	// 23.1 and older:
   113  	//   I220228 16:01:22.487906 18476248525 3@vendor/github.com/cockroachdb/pebble/ingest.go:637 ⋮ [n24,pebble,s24] 33430782  [JOB 10211226] ingested L0:21818678 (1.8 K), L0:21818683 (1.2 K), L0:21818679 (1.6 K), L0:21818680 (1.1 K), L0:21818681 (1.1 K), L0:21818682 (160 M)
   114  	// current:
   115  	//   I220228 16:01:22.487906 18476248525 3@vendor/github.com/cockroachdb/pebble/ingest.go:637 ⋮ [n24,pebble,s24] 33430782  [JOB 10211226] ingested L0:21818678 (1.8KB), L0:21818683 (1.2KB), L0:21818679 (1.6KB), L0:21818680 (1.1KB), L0:21818681 (1.1KB), L0:21818682 (160MB)
   116  	//
   117  	ingestedPattern = regexp.MustCompile(
   118  		`^.*` +
   119  			/* Job ID           */ `\[JOB (?P<job>\d+)]\s` +
   120  			/* ingested */ `ingested\s`)
   121  	ingestedPatternJobIdx = ingestedPattern.SubexpIndex("job")
   122  	ingestedFilePattern   = regexp.MustCompile(
   123  		`L` +
   124  			/* Level       */ `(?P<level>\d):` +
   125  			/* File number */ `(?P<file>\d+)\s` +
   126  			/* Bytes       */ `\((?P<bytes>[0-9.]+( [BKMGTPE]|[KMGTPE]?B))\)`)
   127  	ingestedFilePatternLevelIdx = ingestedFilePattern.SubexpIndex("level")
   128  	ingestedFilePatternFileIdx  = ingestedFilePattern.SubexpIndex("file")
   129  	ingestedFilePatternBytesIdx = ingestedFilePattern.SubexpIndex("bytes")
   130  
   131  	// flushable ingestions
   132  	//
   133  	// I230831 04:13:28.824280 3780 3@pebble/event.go:685 ⋮ [n10,s10,pebble] 365  [JOB 226] flushed 6 ingested flushables L0:024334 (1.5KB) + L0:024339 (1.0KB) + L0:024335 (1.9KB) + L0:024336 (1.1KB) + L0:024337 (1.1KB) + L0:024338 (12KB) in 0.0s (0.0s total), output rate 67MB/s
   134  	flushableIngestedPattern = regexp.MustCompile(
   135  		`^.*` +
   136  			/* Job ID           */ `\[JOB (?P<job>\d+)]\s` +
   137  			/* match ingested flushable */ `flushed \d ingested flushable`)
   138  	flushableIngestedPatternJobIdx = flushableIngestedPattern.SubexpIndex("job")
   139  
   140  	// Example read-amp log line:
   141  	// 23.1 and older:
   142  	//   total     31766   188 G       -   257 G   187 G    48 K   3.6 G     744   536 G    49 K   278 G       5     2.1
   143  	// current:
   144  	//   total |     1   639B     0B |     - |   84B |     0     0B |     0     0B |     3  1.9KB | 1.2KB |   1 23.7
   145  	readAmpPattern = regexp.MustCompile(
   146  		/* Read amp */ `(?:^|\+)(?:\s{2}total|total \|).*?\s(?P<value>\d+)\s.{4,7}$`,
   147  	)
   148  	readAmpPatternValueIdx = readAmpPattern.SubexpIndex("value")
   149  )
   150  
   151  const (
   152  	// timeFmt matches the Cockroach log timestamp format.
   153  	// See: https://github.com/cockroachdb/cockroach/blob/master/pkg/util/log/format_crdb_v2.go
   154  	timeFmt = "060102 15:04:05.000000"
   155  
   156  	// timeFmtSlim is similar to timeFmt, except that it strips components with a
   157  	// lower granularity than a minute.
   158  	timeFmtSlim = "060102 15:04"
   159  
   160  	// timeFmtHrMinSec prints only the hour, minute and second of the time.
   161  	timeFmtHrMinSec = "15:04:05"
   162  )
   163  
   164  // compactionType is the type of compaction. It tracks the types in
   165  // compaction.go. We copy the values here to avoid exporting the types in
   166  // compaction.go.
   167  type compactionType uint8
   168  
   169  const (
   170  	compactionTypeDefault compactionType = iota
   171  	compactionTypeFlush
   172  	compactionTypeMove
   173  	compactionTypeDeleteOnly
   174  	compactionTypeElisionOnly
   175  	compactionTypeRead
   176  )
   177  
   178  // String implements fmt.Stringer.
   179  func (c compactionType) String() string {
   180  	switch c {
   181  	case compactionTypeDefault:
   182  		return "default"
   183  	case compactionTypeMove:
   184  		return "move"
   185  	case compactionTypeDeleteOnly:
   186  		return "delete-only"
   187  	case compactionTypeElisionOnly:
   188  		return "elision-only"
   189  	case compactionTypeRead:
   190  		return "read"
   191  	default:
   192  		panic(errors.Newf("unknown compaction type: %s", c))
   193  	}
   194  }
   195  
   196  // parseCompactionType parses the given compaction type string and returns a
   197  // compactionType.
   198  func parseCompactionType(s string) (t compactionType, err error) {
   199  	switch s {
   200  	case "default":
   201  		t = compactionTypeDefault
   202  	case "move":
   203  		t = compactionTypeMove
   204  	case "delete-only":
   205  		t = compactionTypeDeleteOnly
   206  	case "elision-only":
   207  		t = compactionTypeElisionOnly
   208  	case "read":
   209  		t = compactionTypeRead
   210  	default:
   211  		err = errors.Newf("unknown compaction type: %s", s)
   212  	}
   213  	return
   214  }
   215  
   216  // compactionStart is a compaction start event.
   217  type compactionStart struct {
   218  	ctx        logContext
   219  	jobID      int
   220  	cType      compactionType
   221  	fromLevel  int
   222  	toLevel    int
   223  	inputBytes uint64
   224  }
   225  
   226  // parseCompactionStart converts the given regular expression sub-matches for a
   227  // compaction start log line into a compactionStart event.
   228  func parseCompactionStart(matches []string) (compactionStart, error) {
   229  	var start compactionStart
   230  
   231  	// Parse job ID.
   232  	jobID, err := strconv.Atoi(matches[compactionPatternJobIdx])
   233  	if err != nil {
   234  		return start, errors.Newf("could not parse jobID: %s", err)
   235  	}
   236  
   237  	// Parse compaction type.
   238  	cType, err := parseCompactionType(matches[compactionPatternTypeIdx])
   239  	if err != nil {
   240  		return start, err
   241  	}
   242  
   243  	// Parse input bytes.
   244  	inputBytes, err := sumInputBytes(matches[compactionPatternLevels])
   245  	if err != nil {
   246  		return start, errors.Newf("could not sum input bytes: %s", err)
   247  	}
   248  
   249  	// Parse from-level.
   250  	from, err := strconv.Atoi(matches[compactionPatternFromIdx])
   251  	if err != nil {
   252  		return start, errors.Newf("could not parse from-level: %s", err)
   253  	}
   254  
   255  	// Parse to-level. For deletion and elision compactions, set the same level.
   256  	to := from
   257  	if cType != compactionTypeElisionOnly && cType != compactionTypeDeleteOnly {
   258  		to, err = strconv.Atoi(matches[compactionPatternToIdx])
   259  		if err != nil {
   260  			return start, errors.Newf("could not parse to-level: %s", err)
   261  		}
   262  	}
   263  
   264  	start = compactionStart{
   265  		jobID:      jobID,
   266  		cType:      cType,
   267  		fromLevel:  from,
   268  		toLevel:    to,
   269  		inputBytes: inputBytes,
   270  	}
   271  
   272  	return start, nil
   273  }
   274  
   275  // compactionEnd is a compaction end event.
   276  type compactionEnd struct {
   277  	jobID        int
   278  	writtenBytes uint64
   279  	// TODO(jackson): Parse and include the aggregate size of input
   280  	// sstables. It may be instructive, because compactions that drop
   281  	// keys write less data than they remove from the input level.
   282  }
   283  
   284  // parseCompactionEnd converts the given regular expression sub-matches for a
   285  // compaction end log line into a compactionEnd event.
   286  func parseCompactionEnd(matches []string) (compactionEnd, error) {
   287  	var end compactionEnd
   288  
   289  	// Parse job ID.
   290  	jobID, err := strconv.Atoi(matches[compactionPatternJobIdx])
   291  	if err != nil {
   292  		return end, errors.Newf("could not parse jobID: %s", err)
   293  	}
   294  	end = compactionEnd{jobID: jobID}
   295  
   296  	// Optionally, if we have compacted bytes.
   297  	if matches[compactionPatternBytesIdx] != "" {
   298  		end.writtenBytes = unHumanize(matches[compactionPatternBytesIdx])
   299  	}
   300  
   301  	return end, nil
   302  }
   303  
   304  // parseFlushStart converts the given regular expression sub-matches for a
   305  // memtable flush start log line into a compactionStart event.
   306  func parseFlushStart(matches []string) (compactionStart, error) {
   307  	var start compactionStart
   308  	// Parse job ID.
   309  	jobID, err := strconv.Atoi(matches[flushPatternJobIdx])
   310  	if err != nil {
   311  		return start, errors.Newf("could not parse jobID: %s", err)
   312  	}
   313  	c := compactionStart{
   314  		jobID:     jobID,
   315  		cType:     compactionTypeFlush,
   316  		fromLevel: -1,
   317  		toLevel:   0,
   318  	}
   319  	return c, nil
   320  }
   321  
   322  // parseFlushEnd converts the given regular expression sub-matches for a
   323  // memtable flush end log line into a compactionEnd event.
   324  func parseFlushEnd(matches []string) (compactionEnd, error) {
   325  	var end compactionEnd
   326  
   327  	// Parse job ID.
   328  	jobID, err := strconv.Atoi(matches[flushPatternJobIdx])
   329  	if err != nil {
   330  		return end, errors.Newf("could not parse jobID: %s", err)
   331  	}
   332  	end = compactionEnd{jobID: jobID}
   333  
   334  	// Optionally, if we have flushed bytes.
   335  	if matches[flushPatternBytesIdx] != "" {
   336  		end.writtenBytes = unHumanize(matches[flushPatternBytesIdx])
   337  	}
   338  
   339  	return end, nil
   340  }
   341  
   342  // event describes an aggregated event (eg, start and end events
   343  // combined if necessary).
   344  type event struct {
   345  	nodeID     int
   346  	storeID    int
   347  	jobID      int
   348  	timeStart  time.Time
   349  	timeEnd    time.Time
   350  	compaction *compaction
   351  	ingest     *ingest
   352  }
   353  
   354  // compaction represents an aggregated compaction event (i.e. the combination of
   355  // a start and end event).
   356  type compaction struct {
   357  	cType       compactionType
   358  	fromLevel   int
   359  	toLevel     int
   360  	inputBytes  uint64
   361  	outputBytes uint64
   362  }
   363  
   364  // ingest describes the completion of an ingest.
   365  type ingest struct {
   366  	files []ingestedFile
   367  }
   368  
   369  type ingestedFile struct {
   370  	level     int
   371  	fileNum   int
   372  	sizeBytes uint64
   373  }
   374  
   375  // readAmp represents a read-amp event.
   376  type readAmp struct {
   377  	ctx     logContext
   378  	readAmp int
   379  }
   380  
   381  type nodeStoreJob struct {
   382  	node, store, job int
   383  }
   384  
   385  func (n nodeStoreJob) String() string {
   386  	return fmt.Sprintf("(node=%d,store=%d,job=%d)", n.node, n.store, n.job)
   387  }
   388  
   389  type errorEvent struct {
   390  	path string
   391  	line string
   392  	err  error
   393  }
   394  
   395  // logEventCollector keeps track of open compaction events and read-amp events
   396  // over the course of parsing log line events. Completed compaction events are
   397  // added to the collector once a matching start and end pair are encountered.
   398  // Read-amp events are added as they are encountered (the have no start / end
   399  // concept).
   400  type logEventCollector struct {
   401  	ctx      logContext
   402  	m        map[nodeStoreJob]compactionStart
   403  	events   []event
   404  	readAmps []readAmp
   405  	errors   []errorEvent
   406  }
   407  
   408  // newEventCollector instantiates a new logEventCollector.
   409  func newEventCollector() *logEventCollector {
   410  	return &logEventCollector{
   411  		m: make(map[nodeStoreJob]compactionStart),
   412  	}
   413  }
   414  
   415  // addError records an error encountered during log parsing.
   416  func (c *logEventCollector) addError(path, line string, err error) {
   417  	c.errors = append(c.errors, errorEvent{path: path, line: line, err: err})
   418  }
   419  
   420  // addCompactionStart adds a new compactionStart to the collector. The event is
   421  // tracked by its job ID.
   422  func (c *logEventCollector) addCompactionStart(start compactionStart) error {
   423  	key := nodeStoreJob{c.ctx.node, c.ctx.store, start.jobID}
   424  	if _, ok := c.m[key]; ok {
   425  		return errors.Newf("start event already seen for %s", key)
   426  	}
   427  	start.ctx = c.ctx
   428  	c.m[key] = start
   429  	return nil
   430  }
   431  
   432  // addCompactionEnd completes the compaction event for the given compactionEnd.
   433  func (c *logEventCollector) addCompactionEnd(end compactionEnd) {
   434  	key := nodeStoreJob{c.ctx.node, c.ctx.store, end.jobID}
   435  	start, ok := c.m[key]
   436  	if !ok {
   437  		_, _ = fmt.Fprintf(
   438  			os.Stderr,
   439  			"compaction end event missing start event for %s; skipping\n", key,
   440  		)
   441  		return
   442  	}
   443  
   444  	// Remove the job from the collector once it has been matched.
   445  	delete(c.m, key)
   446  
   447  	c.events = append(c.events, event{
   448  		nodeID:    start.ctx.node,
   449  		storeID:   start.ctx.store,
   450  		jobID:     start.jobID,
   451  		timeStart: start.ctx.timestamp,
   452  		timeEnd:   c.ctx.timestamp,
   453  		compaction: &compaction{
   454  			cType:       start.cType,
   455  			fromLevel:   start.fromLevel,
   456  			toLevel:     start.toLevel,
   457  			inputBytes:  start.inputBytes,
   458  			outputBytes: end.writtenBytes,
   459  		},
   460  	})
   461  }
   462  
   463  // addReadAmp adds the readAmp event to the collector.
   464  func (c *logEventCollector) addReadAmp(ra readAmp) {
   465  	ra.ctx = c.ctx
   466  	c.readAmps = append(c.readAmps, ra)
   467  }
   468  
   469  // logContext captures the metadata of log lines.
   470  type logContext struct {
   471  	timestamp   time.Time
   472  	node, store int
   473  }
   474  
   475  // saveContext saves the given logContext in the collector.
   476  func (c *logEventCollector) saveContext(ctx logContext) {
   477  	c.ctx = ctx
   478  }
   479  
   480  // level is a level in the LSM. The WAL is level -1.
   481  type level int
   482  
   483  // String implements fmt.Stringer.
   484  func (l level) String() string {
   485  	if l == -1 {
   486  		return "WAL"
   487  	}
   488  	return "L" + strconv.Itoa(int(l))
   489  }
   490  
   491  // fromTo is a map key for (from, to) level tuples.
   492  type fromTo struct {
   493  	from, to level
   494  }
   495  
   496  // compactionTypeCount is a mapping from compaction type to count.
   497  type compactionTypeCount map[compactionType]int
   498  
   499  // windowSummary summarizes events in a window of time between a start and end
   500  // time. The window tracks:
   501  // - for each compaction type: counts, total bytes compacted, and total duration.
   502  // - total ingested bytes for each level
   503  // - read amp magnitudes
   504  type windowSummary struct {
   505  	nodeID, storeID      int
   506  	tStart, tEnd         time.Time
   507  	eventCount           int
   508  	flushedCount         int
   509  	flushedBytes         uint64
   510  	flushedTime          time.Duration
   511  	compactionCounts     map[fromTo]compactionTypeCount
   512  	compactionBytesIn    map[fromTo]uint64
   513  	compactionBytesOut   map[fromTo]uint64
   514  	compactionBytesMoved map[fromTo]uint64
   515  	compactionBytesDel   map[fromTo]uint64
   516  	compactionTime       map[fromTo]time.Duration
   517  	ingestedCount        [7]int
   518  	ingestedBytes        [7]uint64
   519  	readAmps             []readAmp
   520  	longRunning          []event
   521  }
   522  
   523  // String implements fmt.Stringer, returning a formatted window summary.
   524  func (s windowSummary) String() string {
   525  	type fromToCount struct {
   526  		ft         fromTo
   527  		counts     compactionTypeCount
   528  		bytesIn    uint64
   529  		bytesOut   uint64
   530  		bytesMoved uint64
   531  		bytesDel   uint64
   532  		duration   time.Duration
   533  	}
   534  	var pairs []fromToCount
   535  	for k, v := range s.compactionCounts {
   536  		pairs = append(pairs, fromToCount{
   537  			ft:         k,
   538  			counts:     v,
   539  			bytesIn:    s.compactionBytesIn[k],
   540  			bytesOut:   s.compactionBytesOut[k],
   541  			bytesMoved: s.compactionBytesMoved[k],
   542  			bytesDel:   s.compactionBytesDel[k],
   543  			duration:   s.compactionTime[k],
   544  		})
   545  	}
   546  	sort.Slice(pairs, func(i, j int) bool {
   547  		l, r := pairs[i], pairs[j]
   548  		if l.ft.from == r.ft.from {
   549  			return l.ft.to < r.ft.to
   550  		}
   551  		return l.ft.from < r.ft.from
   552  	})
   553  
   554  	nodeID, storeID := "?", "?"
   555  	if s.nodeID != -1 {
   556  		nodeID = strconv.Itoa(s.nodeID)
   557  	}
   558  	if s.storeID != -1 {
   559  		storeID = strconv.Itoa(s.storeID)
   560  	}
   561  
   562  	var sb strings.Builder
   563  	sb.WriteString(fmt.Sprintf("node: %s, store: %s\n", nodeID, storeID))
   564  	sb.WriteString(fmt.Sprintf("   from: %s\n", s.tStart.Format(timeFmtSlim)))
   565  	sb.WriteString(fmt.Sprintf("     to: %s\n", s.tEnd.Format(timeFmtSlim)))
   566  	var count, sum int
   567  	for _, ra := range s.readAmps {
   568  		count++
   569  		sum += ra.readAmp
   570  	}
   571  	sb.WriteString(fmt.Sprintf("  r-amp: %.1f\n", float64(sum)/float64(count)))
   572  
   573  	// Print flush+ingest statistics.
   574  	{
   575  		var headerWritten bool
   576  		maybeWriteHeader := func() {
   577  			if !headerWritten {
   578  				sb.WriteString("_kind______from______to_____________________________________count___bytes______time\n")
   579  				headerWritten = true
   580  			}
   581  		}
   582  
   583  		if s.flushedCount > 0 {
   584  			maybeWriteHeader()
   585  			fmt.Fprintf(&sb, "%-7s         %7s                                   %7d %7s %9s\n",
   586  				"flush", "L0", s.flushedCount, humanize.Bytes.Uint64(s.flushedBytes),
   587  				s.flushedTime.Truncate(time.Second))
   588  		}
   589  
   590  		count := s.flushedCount
   591  		sum := s.flushedBytes
   592  		totalTime := s.flushedTime
   593  		for l := 0; l < len(s.ingestedBytes); l++ {
   594  			if s.ingestedCount[l] == 0 {
   595  				continue
   596  			}
   597  			maybeWriteHeader()
   598  			fmt.Fprintf(&sb, "%-7s         %7s                                   %7d %7s\n",
   599  				"ingest", fmt.Sprintf("L%d", l), s.ingestedCount[l], humanize.Bytes.Uint64(s.ingestedBytes[l]))
   600  			count += s.ingestedCount[l]
   601  			sum += s.ingestedBytes[l]
   602  		}
   603  		if headerWritten {
   604  			fmt.Fprintf(&sb, "total                                                     %7d %7s %9s\n",
   605  				count, humanize.Bytes.Uint64(sum), totalTime.Truncate(time.Second),
   606  			)
   607  		}
   608  	}
   609  
   610  	// Print compactions statistics.
   611  	if len(s.compactionCounts) > 0 {
   612  		sb.WriteString("_kind______from______to___default____move___elide__delete___count___in(B)__out(B)__mov(B)__del(B)______time\n")
   613  		var totalDef, totalMove, totalElision, totalDel int
   614  		var totalBytesIn, totalBytesOut, totalBytesMoved, totalBytesDel uint64
   615  		var totalTime time.Duration
   616  		for _, p := range pairs {
   617  			def := p.counts[compactionTypeDefault]
   618  			move := p.counts[compactionTypeMove]
   619  			elision := p.counts[compactionTypeElisionOnly]
   620  			del := p.counts[compactionTypeDeleteOnly]
   621  			total := def + move + elision + del
   622  
   623  			str := fmt.Sprintf("%-7s %7s %7s   %7d %7d %7d %7d %7d %7s %7s %7s %7s %9s\n",
   624  				"compact", p.ft.from, p.ft.to, def, move, elision, del, total,
   625  				humanize.Bytes.Uint64(p.bytesIn), humanize.Bytes.Uint64(p.bytesOut),
   626  				humanize.Bytes.Uint64(p.bytesMoved), humanize.Bytes.Uint64(p.bytesDel),
   627  				p.duration.Truncate(time.Second))
   628  			sb.WriteString(str)
   629  
   630  			totalDef += def
   631  			totalMove += move
   632  			totalElision += elision
   633  			totalDel += del
   634  			totalBytesIn += p.bytesIn
   635  			totalBytesOut += p.bytesOut
   636  			totalBytesMoved += p.bytesMoved
   637  			totalBytesDel += p.bytesDel
   638  			totalTime += p.duration
   639  		}
   640  		sb.WriteString(fmt.Sprintf("total         %19d %7d %7d %7d %7d %7s %7s %7s %7s %9s\n",
   641  			totalDef, totalMove, totalElision, totalDel, s.eventCount,
   642  			humanize.Bytes.Uint64(totalBytesIn), humanize.Bytes.Uint64(totalBytesOut),
   643  			humanize.Bytes.Uint64(totalBytesMoved), humanize.Bytes.Uint64(totalBytesDel),
   644  			totalTime.Truncate(time.Second)))
   645  	}
   646  
   647  	// (Optional) Long running events.
   648  	if len(s.longRunning) > 0 {
   649  		sb.WriteString("long-running events (descending runtime):\n")
   650  		sb.WriteString("_kind________from________to_______job______type_____start_______end____dur(s)_____bytes:\n")
   651  		for _, e := range s.longRunning {
   652  			c := e.compaction
   653  			kind := "compact"
   654  			if c.fromLevel == -1 {
   655  				kind = "flush"
   656  			}
   657  			sb.WriteString(fmt.Sprintf("%-7s %9s %9s %9d %9s %9s %9s %9.0f %9s\n",
   658  				kind, level(c.fromLevel), level(c.toLevel), e.jobID, c.cType,
   659  				e.timeStart.Format(timeFmtHrMinSec), e.timeEnd.Format(timeFmtHrMinSec),
   660  				e.timeEnd.Sub(e.timeStart).Seconds(), humanize.Bytes.Uint64(c.outputBytes)))
   661  		}
   662  	}
   663  
   664  	return sb.String()
   665  }
   666  
   667  // windowSummarySlice is a slice of windowSummary that sorts in order of start
   668  // time, node, then store.
   669  type windowsSummarySlice []windowSummary
   670  
   671  func (s windowsSummarySlice) Len() int {
   672  	return len(s)
   673  }
   674  
   675  func (s windowsSummarySlice) Less(i, j int) bool {
   676  	if !s[i].tStart.Equal(s[j].tStart) {
   677  		return s[i].tStart.Before(s[j].tStart)
   678  	}
   679  	if s[i].nodeID != s[j].nodeID {
   680  		return s[i].nodeID < s[j].nodeID
   681  	}
   682  	return s[i].storeID < s[j].storeID
   683  }
   684  
   685  func (s windowsSummarySlice) Swap(i, j int) {
   686  	s[i], s[j] = s[j], s[i]
   687  }
   688  
   689  // eventSlice is a slice of events that sorts in order of node, store,
   690  // then event start time.
   691  type eventSlice []event
   692  
   693  func (s eventSlice) Len() int {
   694  	return len(s)
   695  }
   696  
   697  func (s eventSlice) Less(i, j int) bool {
   698  	if s[i].nodeID != s[j].nodeID {
   699  		return s[i].nodeID < s[j].nodeID
   700  	}
   701  	if s[i].storeID != s[j].storeID {
   702  		return s[i].storeID < s[j].storeID
   703  	}
   704  	return s[i].timeStart.Before(s[j].timeStart)
   705  }
   706  
   707  func (s eventSlice) Swap(i, j int) {
   708  	s[i], s[j] = s[j], s[i]
   709  }
   710  
   711  // readAmpSlice is a slice of readAmp events that sorts in order of node, store,
   712  // then read amp event start time.
   713  type readAmpSlice []readAmp
   714  
   715  func (r readAmpSlice) Len() int {
   716  	return len(r)
   717  }
   718  
   719  func (r readAmpSlice) Less(i, j int) bool {
   720  	// Sort by node, store, then read-amp.
   721  	if r[i].ctx.node != r[j].ctx.node {
   722  		return r[i].ctx.node < r[j].ctx.node
   723  	}
   724  	if r[i].ctx.store != r[j].ctx.store {
   725  		return r[i].ctx.store < r[j].ctx.store
   726  	}
   727  	return r[i].ctx.timestamp.Before(r[j].ctx.timestamp)
   728  }
   729  
   730  func (r readAmpSlice) Swap(i, j int) {
   731  	r[i], r[j] = r[j], r[i]
   732  }
   733  
   734  // aggregator combines compaction and read-amp events within windows of fixed
   735  // duration and returns one aggregated windowSummary struct per window.
   736  type aggregator struct {
   737  	window           time.Duration
   738  	events           []event
   739  	readAmps         []readAmp
   740  	longRunningLimit time.Duration
   741  }
   742  
   743  // newAggregator returns a new aggregator.
   744  func newAggregator(
   745  	window, longRunningLimit time.Duration, events []event, readAmps []readAmp,
   746  ) *aggregator {
   747  	return &aggregator{
   748  		window:           window,
   749  		events:           events,
   750  		readAmps:         readAmps,
   751  		longRunningLimit: longRunningLimit,
   752  	}
   753  }
   754  
   755  // aggregate aggregates the events into windows, returning the windowSummary for
   756  // each interval.
   757  func (a *aggregator) aggregate() []windowSummary {
   758  	if len(a.events) == 0 {
   759  		return nil
   760  	}
   761  
   762  	// Sort the event and read-amp slices by start time.
   763  	sort.Sort(eventSlice(a.events))
   764  	sort.Sort(readAmpSlice(a.readAmps))
   765  
   766  	initWindow := func(e event) *windowSummary {
   767  		start := e.timeStart.Truncate(a.window)
   768  		return &windowSummary{
   769  			nodeID:               e.nodeID,
   770  			storeID:              e.storeID,
   771  			tStart:               start,
   772  			tEnd:                 start.Add(a.window),
   773  			compactionCounts:     make(map[fromTo]compactionTypeCount),
   774  			compactionBytesIn:    make(map[fromTo]uint64),
   775  			compactionBytesOut:   make(map[fromTo]uint64),
   776  			compactionBytesMoved: make(map[fromTo]uint64),
   777  			compactionBytesDel:   make(map[fromTo]uint64),
   778  			compactionTime:       make(map[fromTo]time.Duration),
   779  		}
   780  	}
   781  
   782  	var windows []windowSummary
   783  	var j int // index for read-amps
   784  	finishWindow := func(cur *windowSummary) {
   785  		// Collect read-amp values for the previous window.
   786  		var readAmps []readAmp
   787  		for j < len(a.readAmps) {
   788  			ra := a.readAmps[j]
   789  
   790  			// Skip values before the current window.
   791  			if ra.ctx.node < cur.nodeID ||
   792  				ra.ctx.store < cur.storeID ||
   793  				ra.ctx.timestamp.Before(cur.tStart) {
   794  				j++
   795  				continue
   796  			}
   797  
   798  			// We've passed over the current window. Stop.
   799  			if ra.ctx.node > cur.nodeID ||
   800  				ra.ctx.store > cur.storeID ||
   801  				ra.ctx.timestamp.After(cur.tEnd) {
   802  				break
   803  			}
   804  
   805  			// Collect this read-amp value.
   806  			readAmps = append(readAmps, ra)
   807  			j++
   808  		}
   809  		cur.readAmps = readAmps
   810  
   811  		// Sort long running compactions in descending order of duration.
   812  		sort.Slice(cur.longRunning, func(i, j int) bool {
   813  			l := cur.longRunning[i]
   814  			r := cur.longRunning[j]
   815  			return l.timeEnd.Sub(l.timeStart) > r.timeEnd.Sub(r.timeStart)
   816  		})
   817  
   818  		// Add the completed window to the set of windows.
   819  		windows = append(windows, *cur)
   820  	}
   821  
   822  	// Move through the compactions, collecting relevant compactions into the same
   823  	// window. Windows have the same node and store, and a compaction start time
   824  	// within a given range.
   825  	i := 0
   826  	curWindow := initWindow(a.events[i])
   827  	for ; ; i++ {
   828  		// No more windows. Complete the current window.
   829  		if i == len(a.events) {
   830  			finishWindow(curWindow)
   831  			break
   832  		}
   833  		e := a.events[i]
   834  
   835  		// If we're at the start of a new interval, finalize the current window and
   836  		// start a new one.
   837  		if curWindow.nodeID != e.nodeID ||
   838  			curWindow.storeID != e.storeID ||
   839  			e.timeStart.After(curWindow.tEnd) {
   840  			finishWindow(curWindow)
   841  			curWindow = initWindow(e)
   842  		}
   843  
   844  		switch {
   845  		case e.ingest != nil:
   846  			// Update ingest stats.
   847  			for _, f := range e.ingest.files {
   848  				curWindow.ingestedCount[f.level]++
   849  				curWindow.ingestedBytes[f.level] += f.sizeBytes
   850  			}
   851  		case e.compaction != nil && e.compaction.cType == compactionTypeFlush:
   852  			// Update flush stats.
   853  			f := e.compaction
   854  			curWindow.flushedCount++
   855  			curWindow.flushedBytes += f.outputBytes
   856  			curWindow.flushedTime += e.timeEnd.Sub(e.timeStart)
   857  		case e.compaction != nil:
   858  			// Update compaction stats.
   859  			c := e.compaction
   860  			// Update compaction counts.
   861  			ft := fromTo{level(c.fromLevel), level(c.toLevel)}
   862  			m, ok := curWindow.compactionCounts[ft]
   863  			if !ok {
   864  				m = make(compactionTypeCount)
   865  				curWindow.compactionCounts[ft] = m
   866  			}
   867  			m[c.cType]++
   868  			curWindow.eventCount++
   869  
   870  			// Update compacted bytes in / out / moved / deleted.
   871  			switch c.cType {
   872  			case compactionTypeMove:
   873  				curWindow.compactionBytesMoved[ft] += c.inputBytes
   874  			case compactionTypeDeleteOnly:
   875  				curWindow.compactionBytesDel[ft] += c.inputBytes
   876  			default:
   877  				curWindow.compactionBytesIn[ft] += c.inputBytes
   878  				curWindow.compactionBytesOut[ft] += c.outputBytes
   879  			}
   880  
   881  			// Update compaction time.
   882  			_, ok = curWindow.compactionTime[ft]
   883  			if !ok {
   884  				curWindow.compactionTime[ft] = 0
   885  			}
   886  			curWindow.compactionTime[ft] += e.timeEnd.Sub(e.timeStart)
   887  
   888  		}
   889  		// Add "long-running" events. Those that start in this window
   890  		// that have duration longer than the window interval.
   891  		if e.timeEnd.Sub(e.timeStart) > a.longRunningLimit {
   892  			curWindow.longRunning = append(curWindow.longRunning, e)
   893  		}
   894  	}
   895  
   896  	// Windows are added in order of (node, store, time). Re-sort the windows by
   897  	// (time, node, store) for better presentation.
   898  	sort.Sort(windowsSummarySlice(windows))
   899  
   900  	return windows
   901  }
   902  
   903  // parseLog parses the log file with the given path, using the given parse
   904  // function to collect events in the given logEventCollector. parseLog
   905  // returns a non-nil error if an I/O error was encountered while reading
   906  // the log file. Parsing errors are accumulated in the
   907  // logEventCollector.
   908  func parseLog(path string, b *logEventCollector) error {
   909  	f, err := os.Open(path)
   910  	if err != nil {
   911  		return err
   912  	}
   913  	defer f.Close()
   914  
   915  	s := bufio.NewScanner(f)
   916  	for s.Scan() {
   917  		line := s.Text()
   918  		// Store the log context for the current line, if we have one.
   919  		if err := parseLogContext(line, b); err != nil {
   920  			return err
   921  		}
   922  
   923  		// First check for a flush or compaction.
   924  		matches := sentinelPattern.FindStringSubmatch(line)
   925  		if matches != nil {
   926  			// Determine which regexp to apply by testing the first letter of the prefix.
   927  			var err error
   928  			switch matches[sentinelPatternPrefixIdx][0] {
   929  			case 'c':
   930  				err = parseCompaction(line, b)
   931  			case 'f':
   932  				err = parseFlush(line, b)
   933  			case 'i':
   934  				err = parseIngest(line, b)
   935  			default:
   936  				err = errors.Newf("unexpected line: neither compaction nor flush: %s", line)
   937  			}
   938  			if err != nil {
   939  				b.addError(path, line, err)
   940  			}
   941  			continue
   942  		}
   943  
   944  		// Else check for an LSM debug line.
   945  		if err = parseReadAmp(line, b); err != nil {
   946  			b.addError(path, line, err)
   947  			continue
   948  		}
   949  	}
   950  	return s.Err()
   951  }
   952  
   953  // parseLogContext extracts contextual information from the log line (e.g. the
   954  // timestamp, node and store).
   955  func parseLogContext(line string, b *logEventCollector) error {
   956  	matches := logContextPattern.FindStringSubmatch(line)
   957  	if matches == nil {
   958  		return nil
   959  	}
   960  
   961  	// Parse start time.
   962  	t, err := time.Parse(timeFmt, matches[logContextPatternTimestampIdx])
   963  	if err != nil {
   964  		return errors.Newf("could not parse timestamp: %s", err)
   965  	}
   966  
   967  	// Parse node and store.
   968  	nodeID, err := strconv.Atoi(matches[logContextPatternNodeIdx])
   969  	if err != nil {
   970  		if matches[logContextPatternNodeIdx] != "?" {
   971  			return errors.Newf("could not parse node ID: %s", err)
   972  		}
   973  		nodeID = -1
   974  	}
   975  
   976  	storeID, err := strconv.Atoi(matches[logContextPatternStoreIdx])
   977  	if err != nil {
   978  		if matches[logContextPatternStoreIdx] != "?" {
   979  			return errors.Newf("could not parse store ID: %s", err)
   980  		}
   981  		storeID = -1
   982  	}
   983  
   984  	b.saveContext(logContext{
   985  		timestamp: t,
   986  		node:      nodeID,
   987  		store:     storeID,
   988  	})
   989  	return nil
   990  }
   991  
   992  // parseCompaction parses and collects Pebble compaction events.
   993  func parseCompaction(line string, b *logEventCollector) error {
   994  	matches := compactionPattern.FindStringSubmatch(line)
   995  	if matches == nil {
   996  		return nil
   997  	}
   998  
   999  	// "compacting": implies start line.
  1000  	if matches[compactionPatternSuffixIdx] == "ing" {
  1001  		start, err := parseCompactionStart(matches)
  1002  		if err != nil {
  1003  			return err
  1004  		}
  1005  		if err := b.addCompactionStart(start); err != nil {
  1006  			return err
  1007  		}
  1008  		return nil
  1009  	}
  1010  
  1011  	// "compacted": implies end line.
  1012  	end, err := parseCompactionEnd(matches)
  1013  	if err != nil {
  1014  		return err
  1015  	}
  1016  
  1017  	b.addCompactionEnd(end)
  1018  	return nil
  1019  }
  1020  
  1021  // parseFlush parses and collects Pebble memtable flush events.
  1022  func parseFlush(line string, b *logEventCollector) error {
  1023  	matches := flushPattern.FindStringSubmatch(line)
  1024  	if matches == nil {
  1025  		return nil
  1026  	}
  1027  
  1028  	if matches[flushPatternSuffixIdx] == "ing" {
  1029  		start, err := parseFlushStart(matches)
  1030  		if err != nil {
  1031  			return err
  1032  		}
  1033  		return b.addCompactionStart(start)
  1034  	}
  1035  
  1036  	end, err := parseFlushEnd(matches)
  1037  	if err != nil {
  1038  		return err
  1039  	}
  1040  
  1041  	b.addCompactionEnd(end)
  1042  	return nil
  1043  }
  1044  
  1045  func parseIngestDuringFlush(line string, b *logEventCollector) error {
  1046  	matches := flushableIngestedPattern.FindStringSubmatch(line)
  1047  	if matches == nil {
  1048  		return nil
  1049  	}
  1050  	// Parse job ID.
  1051  	jobID, err := strconv.Atoi(matches[flushableIngestedPatternJobIdx])
  1052  	if err != nil {
  1053  		return errors.Newf("could not parse jobID: %s", err)
  1054  	}
  1055  	return parseRemainingIngestLogLine(jobID, line, b)
  1056  }
  1057  
  1058  // parseIngest parses and collects Pebble ingest complete events.
  1059  func parseIngest(line string, b *logEventCollector) error {
  1060  	matches := ingestedPattern.FindStringSubmatch(line)
  1061  	if matches == nil {
  1062  		// Try and parse the other kind of ingest.
  1063  		return parseIngestDuringFlush(line, b)
  1064  	}
  1065  	// Parse job ID.
  1066  	jobID, err := strconv.Atoi(matches[ingestedPatternJobIdx])
  1067  	if err != nil {
  1068  		return errors.Newf("could not parse jobID: %s", err)
  1069  	}
  1070  	return parseRemainingIngestLogLine(jobID, line, b)
  1071  }
  1072  
  1073  // parses the level, filenum, and bytes for the files which were ingested.
  1074  func parseRemainingIngestLogLine(jobID int, line string, b *logEventCollector) error {
  1075  	fileMatches := ingestedFilePattern.FindAllStringSubmatch(line, -1)
  1076  	files := make([]ingestedFile, len(fileMatches))
  1077  	for i := range fileMatches {
  1078  		level, err := strconv.Atoi(fileMatches[i][ingestedFilePatternLevelIdx])
  1079  		if err != nil {
  1080  			return errors.Newf("could not parse level: %s", err)
  1081  		}
  1082  		fileNum, err := strconv.Atoi(fileMatches[i][ingestedFilePatternFileIdx])
  1083  		if err != nil {
  1084  			return errors.Newf("could not parse file number: %s", err)
  1085  		}
  1086  		files[i] = ingestedFile{
  1087  			level:     level,
  1088  			fileNum:   fileNum,
  1089  			sizeBytes: unHumanize(fileMatches[i][ingestedFilePatternBytesIdx]),
  1090  		}
  1091  	}
  1092  	b.events = append(b.events, event{
  1093  		nodeID:    b.ctx.node,
  1094  		storeID:   b.ctx.store,
  1095  		jobID:     jobID,
  1096  		timeStart: b.ctx.timestamp,
  1097  		timeEnd:   b.ctx.timestamp,
  1098  		ingest: &ingest{
  1099  			files: files,
  1100  		},
  1101  	})
  1102  	return nil
  1103  }
  1104  
  1105  // parseReadAmp attempts to parse the current line as a read amp value
  1106  func parseReadAmp(line string, b *logEventCollector) error {
  1107  	matches := readAmpPattern.FindStringSubmatch(line)
  1108  	if matches == nil {
  1109  		return nil
  1110  	}
  1111  	val, err := strconv.Atoi(matches[readAmpPatternValueIdx])
  1112  	if err != nil {
  1113  		return errors.Newf("could not parse read amp: %s", err)
  1114  	}
  1115  	b.addReadAmp(readAmp{
  1116  		readAmp: val,
  1117  	})
  1118  	return nil
  1119  }
  1120  
  1121  // runCompactionLogs is runnable function of the top-level cobra.Command that
  1122  // parses and collects Pebble compaction events and LSM information.
  1123  func runCompactionLogs(cmd *cobra.Command, args []string) error {
  1124  	// The args contain a list of log files to read.
  1125  	files := args
  1126  
  1127  	// Scan the log files collecting start and end compaction lines.
  1128  	b := newEventCollector()
  1129  	for _, file := range files {
  1130  		err := parseLog(file, b)
  1131  		// parseLog returns an error only on I/O errors, which we
  1132  		// immediately exit with.
  1133  		if err != nil {
  1134  			return err
  1135  		}
  1136  	}
  1137  
  1138  	window, err := cmd.Flags().GetDuration("window")
  1139  	if err != nil {
  1140  		return err
  1141  	}
  1142  
  1143  	longRunningLimit, err := cmd.Flags().GetDuration("long-running-limit")
  1144  	if err != nil {
  1145  		return err
  1146  	}
  1147  	if longRunningLimit == 0 {
  1148  		// Off by default. Set to infinite duration.
  1149  		longRunningLimit = time.Duration(math.MaxInt64)
  1150  	}
  1151  
  1152  	// Aggregate the lines.
  1153  	a := newAggregator(window, longRunningLimit, b.events, b.readAmps)
  1154  	summaries := a.aggregate()
  1155  	for _, s := range summaries {
  1156  		fmt.Printf("%s\n", s)
  1157  	}
  1158  
  1159  	// After the summaries, print accumulated parsing errors to stderr.
  1160  	for _, e := range b.errors {
  1161  		fmt.Fprintf(os.Stderr, "-\n%s: %s\nError: %s\n", filepath.Base(e.path), e.line, e.err)
  1162  	}
  1163  	return nil
  1164  }
  1165  
  1166  // unHumanize performs the opposite of humanize.Bytes.Uint64 (e.g. "10B",
  1167  // "10MB") or the 23.1 humanize.IEC.Uint64 (e.g. "10 B", "10 M"), converting a
  1168  // human-readable value into a raw number of bytes.
  1169  func unHumanize(s string) uint64 {
  1170  	if len(s) < 2 || !(s[0] >= '0' && s[0] <= '9') {
  1171  		panic(errors.Newf("invalid bytes value %q", s))
  1172  	}
  1173  	if s[len(s)-1] == 'B' {
  1174  		s = s[:len(s)-1]
  1175  	}
  1176  
  1177  	multiplier := uint64(1)
  1178  	switch s[len(s)-1] {
  1179  	case 'K':
  1180  		multiplier = 1 << 10
  1181  	case 'M':
  1182  		multiplier = 1 << 20
  1183  	case 'G':
  1184  		multiplier = 1 << 30
  1185  	case 'T':
  1186  		multiplier = 1 << 40
  1187  	case 'P':
  1188  		multiplier = 1 << 50
  1189  	case 'E':
  1190  		multiplier = 1 << 60
  1191  	}
  1192  	if multiplier != 1 {
  1193  		s = s[:len(s)-1]
  1194  	}
  1195  	if s[len(s)-1] == ' ' {
  1196  		s = s[:len(s)-1]
  1197  	}
  1198  	val, err := strconv.ParseFloat(s, 64)
  1199  	if err != nil {
  1200  		panic(fmt.Sprintf("parsing %s: %v", s, err))
  1201  	}
  1202  
  1203  	return uint64(val * float64(multiplier))
  1204  }
  1205  
  1206  // sumInputBytes takes a string as input and returns the sum of the
  1207  // human-readable sizes, as an integer number of bytes.
  1208  func sumInputBytes(s string) (total uint64, _ error) {
  1209  	var (
  1210  		open bool
  1211  		b    bytes.Buffer
  1212  	)
  1213  	for _, c := range s {
  1214  		switch c {
  1215  		case '(':
  1216  			open = true
  1217  		case ')':
  1218  			total += unHumanize(b.String())
  1219  			b.Reset()
  1220  			open = false
  1221  		default:
  1222  			if open {
  1223  				b.WriteRune(c)
  1224  			}
  1225  		}
  1226  	}
  1227  	return
  1228  }