github.com/zuoyebang/bitalostable@v1.0.1-0.20240229032404-e3b99a834294/record/record_test.go (about)

     1  // Copyright 2011 The LevelDB-Go and Pebble Authors. All rights reserved. Use
     2  // of this source code is governed by a BSD-style license that can be found in
     3  // the LICENSE file.
     4  
     5  package record
     6  
     7  import (
     8  	"bytes"
     9  	"encoding/binary"
    10  	"fmt"
    11  	"io"
    12  	"io/ioutil"
    13  	"math"
    14  	"os"
    15  	"strings"
    16  	"testing"
    17  	"time"
    18  
    19  	"github.com/cockroachdb/errors"
    20  	"github.com/stretchr/testify/require"
    21  	"github.com/zuoyebang/bitalostable/internal/base"
    22  	"golang.org/x/exp/rand"
    23  )
    24  
    25  func short(s string) string {
    26  	if len(s) < 64 {
    27  		return s
    28  	}
    29  	return fmt.Sprintf("%s...(skipping %d bytes)...%s", s[:20], len(s)-40, s[len(s)-20:])
    30  }
    31  
    32  // big returns a string of length n, composed of repetitions of partial.
    33  func big(partial string, n int) string {
    34  	return strings.Repeat(partial, n/len(partial)+1)[:n]
    35  }
    36  
    37  type recordWriter interface {
    38  	WriteRecord([]byte) (int64, error)
    39  	Close() error
    40  }
    41  
    42  func testGeneratorWriter(
    43  	t *testing.T, reset func(), gen func() (string, bool), newWriter func(io.Writer) recordWriter,
    44  ) {
    45  	buf := new(bytes.Buffer)
    46  
    47  	reset()
    48  	w := newWriter(buf)
    49  	for {
    50  		s, ok := gen()
    51  		if !ok {
    52  			break
    53  		}
    54  		if _, err := w.WriteRecord([]byte(s)); err != nil {
    55  			t.Fatalf("Write: %v", err)
    56  		}
    57  	}
    58  	if err := w.Close(); err != nil {
    59  		t.Fatalf("Close: %v", err)
    60  	}
    61  	reset()
    62  	r := NewReader(buf, 0 /* logNum */)
    63  	for {
    64  		s, ok := gen()
    65  		if !ok {
    66  			break
    67  		}
    68  		rr, err := r.Next()
    69  		if err != nil {
    70  			t.Fatalf("reader.Next: %v", err)
    71  		}
    72  		x, err := ioutil.ReadAll(rr)
    73  		if err != nil {
    74  			t.Fatalf("ReadAll: %v", err)
    75  		}
    76  		if string(x) != s {
    77  			t.Fatalf("got %q, want %q", short(string(x)), short(s))
    78  		}
    79  	}
    80  	if _, err := r.Next(); err != io.EOF {
    81  		t.Fatalf("got %v, want %v", err, io.EOF)
    82  	}
    83  }
    84  
    85  func testGenerator(t *testing.T, reset func(), gen func() (string, bool)) {
    86  	t.Run("Writer", func(t *testing.T) {
    87  		testGeneratorWriter(t, reset, gen, func(w io.Writer) recordWriter {
    88  			return NewWriter(w)
    89  		})
    90  	})
    91  
    92  	t.Run("LogWriter", func(t *testing.T) {
    93  		testGeneratorWriter(t, reset, gen, func(w io.Writer) recordWriter {
    94  			return NewLogWriter(w, 0 /* logNum */)
    95  		})
    96  	})
    97  }
    98  
    99  func testLiterals(t *testing.T, s []string) {
   100  	var i int
   101  	reset := func() {
   102  		i = 0
   103  	}
   104  	gen := func() (string, bool) {
   105  		if i == len(s) {
   106  			return "", false
   107  		}
   108  		i++
   109  		return s[i-1], true
   110  	}
   111  	testGenerator(t, reset, gen)
   112  }
   113  
   114  func TestMany(t *testing.T) {
   115  	const n = 1e5
   116  	var i int
   117  	reset := func() {
   118  		i = 0
   119  	}
   120  	gen := func() (string, bool) {
   121  		if i == n {
   122  			return "", false
   123  		}
   124  		i++
   125  		return fmt.Sprintf("%d.", i-1), true
   126  	}
   127  	testGenerator(t, reset, gen)
   128  }
   129  
   130  func TestRandom(t *testing.T) {
   131  	const n = 1e2
   132  	var (
   133  		i int
   134  		r *rand.Rand
   135  	)
   136  	reset := func() {
   137  		i, r = 0, rand.New(rand.NewSource(0))
   138  	}
   139  	gen := func() (string, bool) {
   140  		if i == n {
   141  			return "", false
   142  		}
   143  		i++
   144  		return strings.Repeat(string(uint8(i)), r.Intn(2*blockSize+16)), true
   145  	}
   146  	testGenerator(t, reset, gen)
   147  }
   148  
   149  func TestBasic(t *testing.T) {
   150  	testLiterals(t, []string{
   151  		strings.Repeat("a", 1000),
   152  		strings.Repeat("b", 97270),
   153  		strings.Repeat("c", 8000),
   154  	})
   155  }
   156  
   157  func TestBoundary(t *testing.T) {
   158  	for i := blockSize - 16; i < blockSize+16; i++ {
   159  		s0 := big("abcd", i)
   160  		for j := blockSize - 16; j < blockSize+16; j++ {
   161  			s1 := big("ABCDE", j)
   162  			testLiterals(t, []string{s0, s1})
   163  			testLiterals(t, []string{s0, "", s1})
   164  			testLiterals(t, []string{s0, "x", s1})
   165  		}
   166  	}
   167  }
   168  
   169  func TestFlush(t *testing.T) {
   170  	buf := new(bytes.Buffer)
   171  	w := NewWriter(buf)
   172  	// Write a couple of records. Everything should still be held
   173  	// in the record.Writer buffer, so that buf.Len should be 0.
   174  	w0, _ := w.Next()
   175  	w0.Write([]byte("0"))
   176  	w1, _ := w.Next()
   177  	w1.Write([]byte("11"))
   178  	if got, want := buf.Len(), 0; got != want {
   179  		t.Fatalf("buffer length #0: got %d want %d", got, want)
   180  	}
   181  	// Flush the record.Writer buffer, which should yield 17 bytes.
   182  	// 17 = 2*7 + 1 + 2, which is two headers and 1 + 2 payload bytes.
   183  	require.NoError(t, w.Flush())
   184  	if got, want := buf.Len(), 17; got != want {
   185  		t.Fatalf("buffer length #1: got %d want %d", got, want)
   186  	}
   187  	// Do another write, one that isn't large enough to complete the block.
   188  	// The write should not have flowed through to buf.
   189  	w2, _ := w.Next()
   190  	w2.Write(bytes.Repeat([]byte("2"), 10000))
   191  	if got, want := buf.Len(), 17; got != want {
   192  		t.Fatalf("buffer length #2: got %d want %d", got, want)
   193  	}
   194  	// Flushing should get us up to 10024 bytes written.
   195  	// 10024 = 17 + 7 + 10000.
   196  	require.NoError(t, w.Flush())
   197  	if got, want := buf.Len(), 10024; got != want {
   198  		t.Fatalf("buffer length #3: got %d want %d", got, want)
   199  	}
   200  	// Do a bigger write, one that completes the current block.
   201  	// We should now have 32768 bytes (a complete block), without
   202  	// an explicit flush.
   203  	w3, _ := w.Next()
   204  	w3.Write(bytes.Repeat([]byte("3"), 40000))
   205  	if got, want := buf.Len(), 32768; got != want {
   206  		t.Fatalf("buffer length #4: got %d want %d", got, want)
   207  	}
   208  	// Flushing should get us up to 50038 bytes written.
   209  	// 50038 = 10024 + 2*7 + 40000. There are two headers because
   210  	// the one record was split into two chunks.
   211  	require.NoError(t, w.Flush())
   212  	if got, want := buf.Len(), 50038; got != want {
   213  		t.Fatalf("buffer length #5: got %d want %d", got, want)
   214  	}
   215  	// Check that reading those records give the right lengths.
   216  	r := NewReader(buf, 0 /* logNum */)
   217  	wants := []int64{1, 2, 10000, 40000}
   218  	for i, want := range wants {
   219  		rr, _ := r.Next()
   220  		n, err := io.Copy(ioutil.Discard, rr)
   221  		if err != nil {
   222  			t.Fatalf("read #%d: %v", i, err)
   223  		}
   224  		if n != want {
   225  			t.Fatalf("read #%d: got %d bytes want %d", i, n, want)
   226  		}
   227  	}
   228  }
   229  
   230  func TestNonExhaustiveRead(t *testing.T) {
   231  	const n = 100
   232  	buf := new(bytes.Buffer)
   233  	p := make([]byte, 10)
   234  	rnd := rand.New(rand.NewSource(1))
   235  
   236  	w := NewWriter(buf)
   237  	for i := 0; i < n; i++ {
   238  		length := len(p) + rnd.Intn(3*blockSize)
   239  		s := string(uint8(i)) + "123456789abcdefgh"
   240  		_, _ = w.WriteRecord([]byte(big(s, length)))
   241  	}
   242  	if err := w.Close(); err != nil {
   243  		t.Fatalf("Close: %v", err)
   244  	}
   245  
   246  	r := NewReader(buf, 0 /* logNum */)
   247  	for i := 0; i < n; i++ {
   248  		rr, _ := r.Next()
   249  		_, err := io.ReadFull(rr, p)
   250  		if err != nil {
   251  			t.Fatalf("ReadFull: %v", err)
   252  		}
   253  		want := string(uint8(i)) + "123456789"
   254  		if got := string(p); got != want {
   255  			t.Fatalf("read #%d: got %q want %q", i, got, want)
   256  		}
   257  	}
   258  }
   259  
   260  func TestStaleReader(t *testing.T) {
   261  	buf := new(bytes.Buffer)
   262  
   263  	w := NewWriter(buf)
   264  	_, err := w.WriteRecord([]byte("0"))
   265  	require.NoError(t, err)
   266  
   267  	_, err = w.WriteRecord([]byte("11"))
   268  	require.NoError(t, err)
   269  
   270  	require.NoError(t, w.Close())
   271  
   272  	r := NewReader(buf, 0 /* logNum */)
   273  	r0, err := r.Next()
   274  	require.NoError(t, err)
   275  
   276  	r1, err := r.Next()
   277  	require.NoError(t, err)
   278  
   279  	p := make([]byte, 1)
   280  	if _, err := r0.Read(p); err == nil || !strings.Contains(err.Error(), "stale") {
   281  		t.Fatalf("stale read #0: unexpected error: %v", err)
   282  	}
   283  	if _, err := r1.Read(p); err != nil {
   284  		t.Fatalf("fresh read #1: got %v want nil error", err)
   285  	}
   286  	if p[0] != '1' {
   287  		t.Fatalf("fresh read #1: byte contents: got '%c' want '1'", p[0])
   288  	}
   289  }
   290  
   291  type testRecords struct {
   292  	records [][]byte // The raw value of each record.
   293  	offsets []int64  // The offset of each record within buf, derived from writer.LastRecordOffset.
   294  	buf     []byte   // The serialized records form of all records.
   295  }
   296  
   297  // makeTestRecords generates test records of specified lengths.
   298  // The first record will consist of repeating 0x00 bytes, the next record of
   299  // 0x01 bytes, and so forth. The values will loop back to 0x00 after 0xff.
   300  func makeTestRecords(recordLengths ...int) (*testRecords, error) {
   301  	ret := &testRecords{}
   302  	ret.records = make([][]byte, len(recordLengths))
   303  	ret.offsets = make([]int64, len(recordLengths))
   304  	for i, n := range recordLengths {
   305  		ret.records[i] = bytes.Repeat([]byte{byte(i)}, n)
   306  	}
   307  
   308  	buf := new(bytes.Buffer)
   309  	w := NewWriter(buf)
   310  	for i, rec := range ret.records {
   311  		wRec, err := w.Next()
   312  		if err != nil {
   313  			return nil, err
   314  		}
   315  
   316  		// Alternate between one big write and many small writes.
   317  		cSize := 8
   318  		if i&1 == 0 {
   319  			cSize = len(rec)
   320  		}
   321  		for ; len(rec) > cSize; rec = rec[cSize:] {
   322  			if _, err := wRec.Write(rec[:cSize]); err != nil {
   323  				return nil, err
   324  			}
   325  		}
   326  		if _, err := wRec.Write(rec); err != nil {
   327  			return nil, err
   328  		}
   329  
   330  		ret.offsets[i], err = w.LastRecordOffset()
   331  		if err != nil {
   332  			return nil, err
   333  		}
   334  	}
   335  
   336  	if err := w.Close(); err != nil {
   337  		return nil, err
   338  	}
   339  
   340  	ret.buf = buf.Bytes()
   341  	return ret, nil
   342  }
   343  
   344  // corruptBlock corrupts the checksum of the record that starts at the
   345  // specified block offset. The number of the block offset is 0 based.
   346  func corruptBlock(buf []byte, blockNum int) {
   347  	// Ensure we always permute at least 1 byte of the checksum.
   348  	if buf[blockSize*blockNum] == 0x00 {
   349  		buf[blockSize*blockNum] = 0xff
   350  	} else {
   351  		buf[blockSize*blockNum] = 0x00
   352  	}
   353  
   354  	buf[blockSize*blockNum+1] = 0x00
   355  	buf[blockSize*blockNum+2] = 0x00
   356  	buf[blockSize*blockNum+3] = 0x00
   357  }
   358  
   359  func TestRecoverNoOp(t *testing.T) {
   360  	recs, err := makeTestRecords(
   361  		blockSize-legacyHeaderSize,
   362  		blockSize-legacyHeaderSize,
   363  		blockSize-legacyHeaderSize,
   364  	)
   365  	if err != nil {
   366  		t.Fatalf("makeTestRecords: %v", err)
   367  	}
   368  
   369  	r := NewReader(bytes.NewReader(recs.buf), 0 /* logNum */)
   370  	_, err = r.Next()
   371  	if err != nil || r.err != nil {
   372  		t.Fatalf("reader.Next: %v reader.err: %v", err, r.err)
   373  	}
   374  
   375  	seq, begin, end, n := r.seq, r.begin, r.end, r.n
   376  
   377  	// Should be a no-op since r.err == nil.
   378  	r.recover()
   379  
   380  	// r.err was nil, nothing should have changed.
   381  	if seq != r.seq || begin != r.begin || end != r.end || n != r.n {
   382  		t.Fatal("reader.Recover when no error existed, was not a no-op")
   383  	}
   384  }
   385  
   386  func TestBasicRecover(t *testing.T) {
   387  	recs, err := makeTestRecords(
   388  		blockSize-legacyHeaderSize,
   389  		blockSize-legacyHeaderSize,
   390  		blockSize-legacyHeaderSize,
   391  	)
   392  	if err != nil {
   393  		t.Fatalf("makeTestRecords: %v", err)
   394  	}
   395  
   396  	// Corrupt the checksum of the second record r1 in our file.
   397  	corruptBlock(recs.buf, 1)
   398  
   399  	underlyingReader := bytes.NewReader(recs.buf)
   400  	r := NewReader(underlyingReader, 0 /* logNum */)
   401  
   402  	// The first record r0 should be read just fine.
   403  	r0, err := r.Next()
   404  	if err != nil {
   405  		t.Fatalf("Next: %v", err)
   406  	}
   407  	r0Data, err := ioutil.ReadAll(r0)
   408  	if err != nil {
   409  		t.Fatalf("ReadAll: %v", err)
   410  	}
   411  	if !bytes.Equal(r0Data, recs.records[0]) {
   412  		t.Fatal("Unexpected output in r0's data")
   413  	}
   414  
   415  	// The next record should have a checksum mismatch.
   416  	_, err = r.Next()
   417  	if err == nil {
   418  		t.Fatal("Expected an error while reading a corrupted record")
   419  	}
   420  	if err != ErrInvalidChunk {
   421  		t.Fatalf("Unexpected error returned: %v", err)
   422  	}
   423  
   424  	// Recover from that checksum mismatch.
   425  	r.recover()
   426  	currentOffset, err := underlyingReader.Seek(0, os.SEEK_CUR)
   427  	if err != nil {
   428  		t.Fatalf("current offset: %v", err)
   429  	}
   430  	if currentOffset != blockSize*2 {
   431  		t.Fatalf("current offset: got %d, want %d", currentOffset, blockSize*2)
   432  	}
   433  
   434  	// The third record r2 should be read just fine.
   435  	r2, err := r.Next()
   436  	if err != nil {
   437  		t.Fatalf("Next: %v", err)
   438  	}
   439  	r2Data, err := ioutil.ReadAll(r2)
   440  	if err != nil {
   441  		t.Fatalf("ReadAll: %v", err)
   442  	}
   443  	if !bytes.Equal(r2Data, recs.records[2]) {
   444  		t.Fatal("Unexpected output in r2's data")
   445  	}
   446  }
   447  
   448  func TestRecoverSingleBlock(t *testing.T) {
   449  	// The first record will be blockSize * 3 bytes long. Since each block has
   450  	// a 7 byte header, the first record will roll over into 4 blocks.
   451  	recs, err := makeTestRecords(
   452  		blockSize*3,
   453  		blockSize-legacyHeaderSize,
   454  		blockSize/2,
   455  	)
   456  	if err != nil {
   457  		t.Fatalf("makeTestRecords: %v", err)
   458  	}
   459  
   460  	// Corrupt the checksum for the portion of the first record that exists in
   461  	// the 4th block.
   462  	corruptBlock(recs.buf, 3)
   463  
   464  	// The first record should fail, but only when we read deeper beyond the
   465  	// first block.
   466  	r := NewReader(bytes.NewReader(recs.buf), 0 /* logNum */)
   467  	r0, err := r.Next()
   468  	if err != nil {
   469  		t.Fatalf("Next: %v", err)
   470  	}
   471  
   472  	// Reading deeper should yield a checksum mismatch.
   473  	_, err = ioutil.ReadAll(r0)
   474  	if err == nil {
   475  		t.Fatal("Expected a checksum mismatch error, got nil")
   476  	}
   477  	if err != ErrInvalidChunk {
   478  		t.Fatalf("Unexpected error returned: %v", err)
   479  	}
   480  
   481  	// Recover from that checksum mismatch.
   482  	r.recover()
   483  
   484  	// All of the data in the second record r1 is lost because the first record
   485  	// r0 shared a partial block with it. The second record also overlapped
   486  	// into the block with the third record r2. Recovery should jump to that
   487  	// block, skipping over the end of the second record and start parsing the
   488  	// third record.
   489  	r2, err := r.Next()
   490  	if err != nil {
   491  		t.Fatalf("Next: %v", err)
   492  	}
   493  	r2Data, _ := ioutil.ReadAll(r2)
   494  	if !bytes.Equal(r2Data, recs.records[2]) {
   495  		t.Fatal("Unexpected output in r2's data")
   496  	}
   497  }
   498  
   499  func TestRecoverMultipleBlocks(t *testing.T) {
   500  	recs, err := makeTestRecords(
   501  		// The first record will consume 3 entire blocks but a fraction of the 4th.
   502  		blockSize*3,
   503  		// The second record will completely fill the remainder of the 4th block.
   504  		3*(blockSize-legacyHeaderSize)-2*blockSize-2*legacyHeaderSize,
   505  		// Consume the entirety of the 5th block.
   506  		blockSize-legacyHeaderSize,
   507  		// Consume the entirety of the 6th block.
   508  		blockSize-legacyHeaderSize,
   509  		// Consume roughly half of the 7th block.
   510  		blockSize/2,
   511  	)
   512  	if err != nil {
   513  		t.Fatalf("makeTestRecords: %v", err)
   514  	}
   515  
   516  	// Corrupt the checksum for the portion of the first record that exists in the 4th block.
   517  	corruptBlock(recs.buf, 3)
   518  
   519  	// Now corrupt the two blocks in a row that correspond to recs.records[2:4].
   520  	corruptBlock(recs.buf, 4)
   521  	corruptBlock(recs.buf, 5)
   522  
   523  	// The first record should fail, but only when we read deeper beyond the first block.
   524  	r := NewReader(bytes.NewReader(recs.buf), 0 /* logNum */)
   525  	r0, err := r.Next()
   526  	if err != nil {
   527  		t.Fatalf("Next: %v", err)
   528  	}
   529  
   530  	// Reading deeper should yield a checksum mismatch.
   531  	_, err = ioutil.ReadAll(r0)
   532  	if err == nil {
   533  		t.Fatal("Exptected a checksum mismatch error, got nil")
   534  	}
   535  	if err != ErrInvalidChunk {
   536  		t.Fatalf("Unexpected error returned: %v", err)
   537  	}
   538  
   539  	// Recover from that checksum mismatch.
   540  	r.recover()
   541  
   542  	// All of the data in the second record is lost because the first
   543  	// record shared a partial block with it. The following two records
   544  	// have corrupted checksums as well, so the call above to r.Recover
   545  	// should result in r.Next() being a reader to the 5th record.
   546  	r4, err := r.Next()
   547  	if err != nil {
   548  		t.Fatalf("Next: %v", err)
   549  	}
   550  
   551  	r4Data, _ := ioutil.ReadAll(r4)
   552  	if !bytes.Equal(r4Data, recs.records[4]) {
   553  		t.Fatal("Unexpected output in r4's data")
   554  	}
   555  }
   556  
   557  // verifyLastBlockRecover reads each record from recs expecting that the
   558  // last record will be corrupted. It will then try Recover and verify that EOF
   559  // is returned.
   560  func verifyLastBlockRecover(recs *testRecords) error {
   561  	r := NewReader(bytes.NewReader(recs.buf), 0 /* logNum */)
   562  	// Loop to one element larger than the number of records to verify EOF.
   563  	for i := 0; i < len(recs.records)+1; i++ {
   564  		_, err := r.Next()
   565  		switch i {
   566  		case len(recs.records) - 1:
   567  			if err == nil {
   568  				return errors.New("Expected a checksum mismatch error, got nil")
   569  			}
   570  			r.recover()
   571  		case len(recs.records):
   572  			if err != io.EOF {
   573  				return errors.Errorf("Expected io.EOF, got %v", err)
   574  			}
   575  		default:
   576  			if err != nil {
   577  				return errors.Errorf("Next: %v", err)
   578  			}
   579  		}
   580  	}
   581  	return nil
   582  }
   583  
   584  func TestRecoverLastPartialBlock(t *testing.T) {
   585  	recs, err := makeTestRecords(
   586  		// The first record will consume 3 entire blocks but a fraction of the 4th.
   587  		blockSize*3,
   588  		// The second record will completely fill the remainder of the 4th block.
   589  		3*(blockSize-legacyHeaderSize)-2*blockSize-2*legacyHeaderSize,
   590  		// Consume roughly half of the 5th block.
   591  		blockSize/2,
   592  	)
   593  	if err != nil {
   594  		t.Fatalf("makeTestRecords: %v", err)
   595  	}
   596  
   597  	// Corrupt the 5th block.
   598  	corruptBlock(recs.buf, 4)
   599  
   600  	// Verify Recover works when the last block is corrupted.
   601  	if err := verifyLastBlockRecover(recs); err != nil {
   602  		t.Fatalf("verifyLastBlockRecover: %v", err)
   603  	}
   604  }
   605  
   606  func TestRecoverLastCompleteBlock(t *testing.T) {
   607  	recs, err := makeTestRecords(
   608  		// The first record will consume 3 entire blocks but a fraction of the 4th.
   609  		blockSize*3,
   610  		// The second record will completely fill the remainder of the 4th block.
   611  		3*(blockSize-legacyHeaderSize)-2*blockSize-2*legacyHeaderSize,
   612  		// Consume the entire 5th block.
   613  		blockSize-legacyHeaderSize,
   614  	)
   615  	if err != nil {
   616  		t.Fatalf("makeTestRecords: %v", err)
   617  	}
   618  
   619  	// Corrupt the 5th block.
   620  	corruptBlock(recs.buf, 4)
   621  
   622  	// Verify Recover works when the last block is corrupted.
   623  	if err := verifyLastBlockRecover(recs); err != nil {
   624  		t.Fatalf("verifyLastBlockRecover: %v", err)
   625  	}
   626  }
   627  
   628  func TestReaderOffset(t *testing.T) {
   629  	recs, err := makeTestRecords(
   630  		blockSize*2,
   631  		400,
   632  		500,
   633  		600,
   634  		700,
   635  		800,
   636  		9000,
   637  		1000,
   638  	)
   639  	if err != nil {
   640  		t.Fatalf("makeTestRecords: %v", err)
   641  	}
   642  
   643  	// The first record should fail, but only when we read deeper beyond the first block.
   644  	r := NewReader(bytes.NewReader(recs.buf), 0 /* logNum */)
   645  	for i, offset := range recs.offsets {
   646  		if offset != r.Offset() {
   647  			t.Fatalf("%d: expected offset %d, but found %d", i, offset, r.Offset())
   648  		}
   649  		rec, err := r.Next()
   650  		if err != nil {
   651  			t.Fatalf("Next: %v", err)
   652  		}
   653  		if _, err = ioutil.ReadAll(rec); err != nil {
   654  			t.Fatalf("ReadAll: %v", err)
   655  		}
   656  	}
   657  }
   658  
   659  func TestSeekRecord(t *testing.T) {
   660  	recs, err := makeTestRecords(
   661  		// The first record will consume 3 entire blocks but a fraction of the 4th.
   662  		blockSize*3,
   663  		// The second record will completely fill the remainder of the 4th block.
   664  		3*(blockSize-legacyHeaderSize)-2*blockSize-2*legacyHeaderSize,
   665  		// Consume the entirety of the 5th block.
   666  		blockSize-legacyHeaderSize,
   667  		// Consume the entirety of the 6th block.
   668  		blockSize-legacyHeaderSize,
   669  		// Consume roughly half of the 7th block.
   670  		blockSize/2,
   671  	)
   672  	if err != nil {
   673  		t.Fatalf("makeTestRecords: %v", err)
   674  	}
   675  
   676  	r := NewReader(bytes.NewReader(recs.buf), 0 /* logNum */)
   677  	// Seek to a valid block offset, but within a multiblock record. This should cause the next call to
   678  	// Next after SeekRecord to return the next valid FIRST/FULL chunk of the subsequent record.
   679  	err = r.seekRecord(blockSize)
   680  	if err != nil {
   681  		t.Fatalf("SeekRecord: %v", err)
   682  	}
   683  	rec, err := r.Next()
   684  	if err != nil {
   685  		t.Fatalf("Next: %v", err)
   686  	}
   687  	rData, _ := ioutil.ReadAll(rec)
   688  	if !bytes.Equal(rData, recs.records[1]) {
   689  		t.Fatalf("Unexpected output in record 1's data, got %v want %v", rData, recs.records[1])
   690  	}
   691  
   692  	// Seek 3 bytes into the second block, which is still in the middle of the first record, but not
   693  	// at a valid chunk boundary. Should result in an error upon calling r.Next.
   694  	err = r.seekRecord(blockSize + 3)
   695  	if err != nil {
   696  		t.Fatalf("SeekRecord: %v", err)
   697  	}
   698  	if _, err = r.Next(); err == nil {
   699  		t.Fatalf("Expected an error seeking to an invalid chunk boundary")
   700  	}
   701  	r.recover()
   702  
   703  	// Seek to the fifth block and verify all records can be read as appropriate.
   704  	err = r.seekRecord(blockSize * 4)
   705  	if err != nil {
   706  		t.Fatalf("SeekRecord: %v", err)
   707  	}
   708  
   709  	check := func(i int) {
   710  		for ; i < len(recs.records); i++ {
   711  			rec, err := r.Next()
   712  			if err != nil {
   713  				t.Fatalf("Next: %v", err)
   714  			}
   715  
   716  			rData, _ := ioutil.ReadAll(rec)
   717  			if !bytes.Equal(rData, recs.records[i]) {
   718  				t.Fatalf("Unexpected output in record #%d's data, got %v want %v", i, rData, recs.records[i])
   719  			}
   720  		}
   721  	}
   722  	check(2)
   723  
   724  	// Seek back to the fourth block, and read all subsequent records and verify them.
   725  	err = r.seekRecord(blockSize * 3)
   726  	if err != nil {
   727  		t.Fatalf("SeekRecord: %v", err)
   728  	}
   729  	check(1)
   730  
   731  	// Now seek past the end of the file and verify it causes an error.
   732  	err = r.seekRecord(1 << 20)
   733  	if err == nil {
   734  		t.Fatalf("Seek past the end of a file didn't cause an error")
   735  	}
   736  	if err != io.ErrUnexpectedEOF {
   737  		t.Fatalf("Seeking past EOF raised unexpected error: %v", err)
   738  	}
   739  	r.recover() // Verify recovery works.
   740  
   741  	// Validate the current records are returned after seeking to a valid offset.
   742  	err = r.seekRecord(blockSize * 4)
   743  	if err != nil {
   744  		t.Fatalf("SeekRecord: %v", err)
   745  	}
   746  	check(2)
   747  }
   748  
   749  func TestLastRecordOffset(t *testing.T) {
   750  	recs, err := makeTestRecords(
   751  		// The first record will consume 3 entire blocks but a fraction of the 4th.
   752  		blockSize*3,
   753  		// The second record will completely fill the remainder of the 4th block.
   754  		3*(blockSize-legacyHeaderSize)-2*blockSize-2*legacyHeaderSize,
   755  		// Consume the entirety of the 5th block.
   756  		blockSize-legacyHeaderSize,
   757  		// Consume the entirety of the 6th block.
   758  		blockSize-legacyHeaderSize,
   759  		// Consume roughly half of the 7th block.
   760  		blockSize/2,
   761  	)
   762  	if err != nil {
   763  		t.Fatalf("makeTestRecords: %v", err)
   764  	}
   765  
   766  	wants := []int64{0, 98332, 131072, 163840, 196608}
   767  	for i, got := range recs.offsets {
   768  		if want := wants[i]; got != want {
   769  			t.Errorf("record #%d: got %d, want %d", i, got, want)
   770  		}
   771  	}
   772  }
   773  
   774  func TestNoLastRecordOffset(t *testing.T) {
   775  	buf := new(bytes.Buffer)
   776  	w := NewWriter(buf)
   777  	defer w.Close()
   778  
   779  	if _, err := w.LastRecordOffset(); err != ErrNoLastRecord {
   780  		t.Fatalf("Expected ErrNoLastRecord, got: %v", err)
   781  	}
   782  
   783  	require.NoError(t, w.Flush())
   784  
   785  	if _, err := w.LastRecordOffset(); err != ErrNoLastRecord {
   786  		t.Fatalf("LastRecordOffset: got: %v, want ErrNoLastRecord", err)
   787  	}
   788  
   789  	_, err := w.WriteRecord([]byte("testrecord"))
   790  	require.NoError(t, err)
   791  
   792  	if off, err := w.LastRecordOffset(); err != nil {
   793  		t.Fatalf("LastRecordOffset: %v", err)
   794  	} else if off != 0 {
   795  		t.Fatalf("LastRecordOffset: got %d, want 0", off)
   796  	}
   797  }
   798  
   799  func TestInvalidLogNum(t *testing.T) {
   800  	var buf bytes.Buffer
   801  	w := NewLogWriter(&buf, 1)
   802  	for i := 0; i < 10; i++ {
   803  		s := fmt.Sprintf("%04d\n", i)
   804  		_, err := w.WriteRecord([]byte(s))
   805  		require.NoError(t, err)
   806  	}
   807  	require.NoError(t, w.Close())
   808  
   809  	{
   810  		r := NewReader(bytes.NewReader(buf.Bytes()), 1)
   811  		for i := 0; i < 10; i++ {
   812  			rr, err := r.Next()
   813  			require.NoError(t, err)
   814  
   815  			x, err := ioutil.ReadAll(rr)
   816  			require.NoError(t, err)
   817  
   818  			s := fmt.Sprintf("%04d\n", i)
   819  			if s != string(x) {
   820  				t.Fatalf("expected %s, but found %s", s, x)
   821  			}
   822  		}
   823  		if _, err := r.Next(); err != io.EOF {
   824  			t.Fatalf("expected EOF, but found %s", err)
   825  		}
   826  	}
   827  
   828  	{
   829  		r := NewReader(bytes.NewReader(buf.Bytes()), 2)
   830  		if _, err := r.Next(); err != io.EOF {
   831  			t.Fatalf("expected %s, but found %s\n", io.EOF, err)
   832  		}
   833  	}
   834  }
   835  
   836  func TestSize(t *testing.T) {
   837  	var buf bytes.Buffer
   838  	zeroes := make([]byte, 8<<10)
   839  	w := NewWriter(&buf)
   840  	for i := 0; i < 100; i++ {
   841  		n := rand.Intn(len(zeroes))
   842  		_, err := w.WriteRecord(zeroes[:n])
   843  		require.NoError(t, err)
   844  		require.NoError(t, w.Flush())
   845  		if buf.Len() != int(w.Size()) {
   846  			t.Fatalf("expected %d, but found %d", buf.Len(), w.Size())
   847  		}
   848  	}
   849  	require.NoError(t, w.Close())
   850  }
   851  
   852  type limitedWriter struct {
   853  	io.Writer
   854  	limit int
   855  }
   856  
   857  func (w *limitedWriter) Write(p []byte) (n int, err error) {
   858  	w.limit--
   859  	if w.limit < 0 {
   860  		return len(p), nil
   861  	}
   862  	return w.Writer.Write(p)
   863  }
   864  
   865  func TestRecycleLog(t *testing.T) {
   866  	const min = 16
   867  	const max = 4096
   868  
   869  	rnd := rand.New(rand.NewSource(uint64(time.Now().UnixNano())))
   870  	randBlock := func() []byte {
   871  		data := make([]byte, rand.Intn(max-min)+min)
   872  		tmp := data
   873  		for len(tmp) >= 8 {
   874  			binary.LittleEndian.PutUint64(tmp, rand.Uint64())
   875  			tmp = tmp[8:]
   876  		}
   877  		r := rand.Uint64()
   878  		for i := 0; i < len(tmp); i++ {
   879  			tmp[i] = byte(r)
   880  			r >>= 8
   881  		}
   882  		return data
   883  	}
   884  
   885  	// Recycle a log file 100 times, writing a random number of records filled
   886  	// with random data.
   887  	backing := make([]byte, 1<<20)
   888  	for i := 1; i <= 100; i++ {
   889  		blocks := rnd.Intn(100)
   890  		limitedBuf := &limitedWriter{
   891  			Writer: bytes.NewBuffer(backing[:0]),
   892  			limit:  blocks,
   893  		}
   894  
   895  		w := NewLogWriter(limitedBuf, base.FileNum(i))
   896  		sizes := make([]int, 10+rnd.Intn(100))
   897  		for j := range sizes {
   898  			data := randBlock()
   899  			if _, err := w.WriteRecord(data); err != nil {
   900  				t.Fatalf("%d/%d: %v", i, j, err)
   901  			}
   902  			sizes[j] = len(data)
   903  		}
   904  		if err := w.Close(); err != nil {
   905  			t.Fatalf("%d: %v", i, err)
   906  		}
   907  
   908  		r := NewReader(bytes.NewReader(backing), base.FileNum(i))
   909  		for j := range sizes {
   910  			rr, err := r.Next()
   911  			if err != nil {
   912  				// If we limited output then an EOF, zeroed, or invalid chunk is expected.
   913  				if limitedBuf.limit < 0 && (err == io.EOF || err == ErrZeroedChunk || err == ErrInvalidChunk) {
   914  					break
   915  				}
   916  				t.Fatalf("%d/%d: %v", i, j, err)
   917  			}
   918  			x, err := ioutil.ReadAll(rr)
   919  			if err != nil {
   920  				// If we limited output then an EOF, zeroed, or invalid chunk is expected.
   921  				if limitedBuf.limit < 0 && (err == io.EOF || err == ErrZeroedChunk || err == ErrInvalidChunk) {
   922  					break
   923  				}
   924  				t.Fatalf("%d/%d: %v", i, j, err)
   925  			}
   926  			if sizes[j] != len(x) {
   927  				t.Fatalf("%d/%d: expected record %d, but found %d", i, j, sizes[j], len(x))
   928  			}
   929  		}
   930  		if _, err := r.Next(); err != io.EOF && err != ErrZeroedChunk && err != ErrInvalidChunk {
   931  			t.Fatalf("%d: expected EOF, but found %v", i, err)
   932  		}
   933  	}
   934  }
   935  
   936  func TestTruncatedLog(t *testing.T) {
   937  	backing := make([]byte, 2*blockSize)
   938  	w := NewLogWriter(bytes.NewBuffer(backing[:0]), base.FileNum(1))
   939  	// Write a record that spans 2 blocks.
   940  	_, err := w.WriteRecord(bytes.Repeat([]byte("s"), blockSize+100))
   941  	require.NoError(t, err)
   942  	require.NoError(t, w.Close())
   943  	// Create a reader only for the first block.
   944  	r := NewReader(bytes.NewReader(backing[:blockSize]), base.FileNum(1))
   945  	rr, err := r.Next()
   946  	require.NoError(t, err)
   947  	_, err = ioutil.ReadAll(rr)
   948  	require.EqualValues(t, err, io.ErrUnexpectedEOF)
   949  }
   950  
   951  func TestRecycleLogWithPartialBlock(t *testing.T) {
   952  	backing := make([]byte, 27)
   953  	w := NewLogWriter(bytes.NewBuffer(backing[:0]), base.FileNum(1))
   954  	// Will write a chunk with 11 byte header + 5 byte payload.
   955  	_, err := w.WriteRecord([]byte("aaaaa"))
   956  	require.NoError(t, err)
   957  	// Close will write a 11-byte EOF chunk.
   958  	require.NoError(t, w.Close())
   959  
   960  	w = NewLogWriter(bytes.NewBuffer(backing[:0]), base.FileNum(2))
   961  	// Will write a chunk with 11 byte header + 1 byte payload.
   962  	_, err = w.WriteRecord([]byte("a"))
   963  	require.NoError(t, err)
   964  	// Close will write a 11-byte EOF chunk.
   965  	require.NoError(t, w.Close())
   966  
   967  	r := NewReader(bytes.NewReader(backing), base.FileNum(2))
   968  	_, err = r.Next()
   969  	require.NoError(t, err)
   970  	// 4 bytes left, which are not enough for even the legacy header.
   971  	if _, err = r.Next(); err != io.EOF {
   972  		t.Fatalf("unexpected error: %v", err)
   973  	}
   974  }
   975  
   976  func TestRecycleLogNumberOverflow(t *testing.T) {
   977  	// We truncate log numbers to 32-bits when writing to the WAL. Test log
   978  	// recycling at the wraparound point, ensuring that EOF chunks are
   979  	// interpreted correctly.
   980  
   981  	backing := make([]byte, 27)
   982  	w := NewLogWriter(bytes.NewBuffer(backing[:0]), base.FileNum(math.MaxUint32))
   983  	// Will write a chunk with 11 byte header + 5 byte payload.
   984  	_, err := w.WriteRecord([]byte("aaaaa"))
   985  	require.NoError(t, err)
   986  	// Close will write a 11-byte EOF chunk.
   987  	require.NoError(t, w.Close())
   988  
   989  	w = NewLogWriter(bytes.NewBuffer(backing[:0]), base.FileNum(math.MaxUint32+1))
   990  	// Will write a chunk with 11 byte header + 1 byte payload.
   991  	_, err = w.WriteRecord([]byte("a"))
   992  	require.NoError(t, err)
   993  	// Close will write a 11-byte EOF chunk.
   994  	require.NoError(t, w.Close())
   995  
   996  	r := NewReader(bytes.NewReader(backing), base.FileNum(math.MaxUint32+1))
   997  	_, err = r.Next()
   998  	require.NoError(t, err)
   999  	// 4 bytes left, which are not enough for even the legacy header.
  1000  	if _, err = r.Next(); err != io.EOF {
  1001  		t.Fatalf("unexpected error: %v", err)
  1002  	}
  1003  }
  1004  
  1005  func TestRecycleLogWithPartialRecord(t *testing.T) {
  1006  	const recordSize = (blockSize * 3) / 2
  1007  
  1008  	// Write a record that is larger than the log block size.
  1009  	backing1 := make([]byte, 2*blockSize)
  1010  	w := NewLogWriter(bytes.NewBuffer(backing1[:0]), base.FileNum(1))
  1011  	_, err := w.WriteRecord(bytes.Repeat([]byte("a"), recordSize))
  1012  	require.NoError(t, err)
  1013  	require.NoError(t, w.Close())
  1014  
  1015  	// Write another record to a new incarnation of the WAL that is larger than
  1016  	// the block size.
  1017  	backing2 := make([]byte, 2*blockSize)
  1018  	w = NewLogWriter(bytes.NewBuffer(backing2[:0]), base.FileNum(2))
  1019  	_, err = w.WriteRecord(bytes.Repeat([]byte("b"), recordSize))
  1020  	require.NoError(t, err)
  1021  	require.NoError(t, w.Close())
  1022  
  1023  	// Copy the second block from the first WAL to the second block of the second
  1024  	// WAL. This produces a scenario where it appears we crashed after writing
  1025  	// the first block of the second WAL, but before writing the second block.
  1026  	copy(backing2[blockSize:], backing1[blockSize:])
  1027  
  1028  	// Verify that we can't read a partial record from the second WAL.
  1029  	r := NewReader(bytes.NewReader(backing2), base.FileNum(2))
  1030  	rr, err := r.Next()
  1031  	require.NoError(t, err)
  1032  
  1033  	_, err = ioutil.ReadAll(rr)
  1034  	require.Equal(t, err, ErrInvalidChunk)
  1035  }
  1036  
  1037  func BenchmarkRecordWrite(b *testing.B) {
  1038  	for _, size := range []int{8, 16, 32, 64, 256, 1028, 4096, 65_536} {
  1039  		b.Run(fmt.Sprintf("size=%d", size), func(b *testing.B) {
  1040  			w := NewLogWriter(ioutil.Discard, 0 /* logNum */)
  1041  			defer w.Close()
  1042  			buf := make([]byte, size)
  1043  
  1044  			b.SetBytes(int64(len(buf)))
  1045  			b.ResetTimer()
  1046  			for i := 0; i < b.N; i++ {
  1047  				if _, err := w.WriteRecord(buf); err != nil {
  1048  					b.Fatal(err)
  1049  				}
  1050  			}
  1051  			b.StopTimer()
  1052  		})
  1053  	}
  1054  }