github.com/keybase/client/go@v0.0.0-20241007131713-f10651d043c8/kbfs/libkbfs/block_retrieval_worker_test.go (about)

     1  // Copyright 2016 Keybase Inc. All rights reserved.
     2  // Use of this source code is governed by a BSD
     3  // license that can be found in the LICENSE file.
     4  package libkbfs
     5  
     6  import (
     7  	"errors"
     8  	"sync"
     9  	"testing"
    10  	"time"
    11  
    12  	"github.com/keybase/client/go/kbfs/data"
    13  	"github.com/keybase/client/go/kbfs/env"
    14  	"github.com/keybase/client/go/kbfs/kbfscodec"
    15  	"github.com/keybase/client/go/kbfs/kbfscrypto"
    16  	"github.com/keybase/client/go/kbfs/libkey"
    17  	"github.com/stretchr/testify/require"
    18  	"golang.org/x/net/context"
    19  )
    20  
    21  // blockReturner contains a block value to copy into requested blocks, and a
    22  // channel to synchronize on with the worker.
    23  type blockReturner struct {
    24  	block      data.Block
    25  	continueCh chan error
    26  	startCh    chan struct{}
    27  }
    28  
    29  // fakeBlockGetter allows specifying and obtaining fake blocks.
    30  type fakeBlockGetter struct {
    31  	mtx           sync.RWMutex
    32  	blockMap      map[data.BlockPointer]blockReturner
    33  	codec         kbfscodec.Codec
    34  	respectCancel bool
    35  }
    36  
    37  // newFakeBlockGetter returns a fakeBlockGetter.
    38  func newFakeBlockGetter(respectCancel bool) *fakeBlockGetter {
    39  	return &fakeBlockGetter{
    40  		blockMap:      make(map[data.BlockPointer]blockReturner),
    41  		codec:         kbfscodec.NewMsgpack(),
    42  		respectCancel: respectCancel,
    43  	}
    44  }
    45  
    46  // setBlockToReturn sets the block that will be returned for a given
    47  // BlockPointer. Returns a writeable channel that getBlock will wait on, to
    48  // allow synchronization of tests.
    49  func (bg *fakeBlockGetter) setBlockToReturn(blockPtr data.BlockPointer,
    50  	block data.Block) (startCh <-chan struct{}, continueCh chan<- error) {
    51  	bg.mtx.Lock()
    52  	defer bg.mtx.Unlock()
    53  	sCh, cCh := make(chan struct{}), make(chan error)
    54  	bg.blockMap[blockPtr] = blockReturner{
    55  		block:      block,
    56  		startCh:    sCh,
    57  		continueCh: cCh,
    58  	}
    59  	return sCh, cCh
    60  }
    61  
    62  // getBlock implements the interface for realBlockGetter.
    63  func (bg *fakeBlockGetter) getBlock(
    64  	ctx context.Context, kmd libkey.KeyMetadata, blockPtr data.BlockPointer,
    65  	block data.Block, _ DiskBlockCacheType) error {
    66  	bg.mtx.RLock()
    67  	defer bg.mtx.RUnlock()
    68  	source, ok := bg.blockMap[blockPtr]
    69  	if !ok {
    70  		return errors.New("Block doesn't exist in fake block map")
    71  	}
    72  	cancelCh := make(chan struct{})
    73  	if bg.respectCancel {
    74  		go func() {
    75  			<-ctx.Done()
    76  			close(cancelCh)
    77  		}()
    78  	}
    79  	// Wait until the caller tells us to continue
    80  	for {
    81  		select {
    82  		case source.startCh <- struct{}{}:
    83  		case err := <-source.continueCh:
    84  			if err != nil {
    85  				return err
    86  			}
    87  			block.Set(source.block)
    88  			return nil
    89  		case <-cancelCh:
    90  			return ctx.Err()
    91  		}
    92  	}
    93  }
    94  
    95  func (bg *fakeBlockGetter) assembleBlock(ctx context.Context,
    96  	kmd libkey.KeyMetadata, ptr data.BlockPointer, block data.Block, buf []byte,
    97  	serverHalf kbfscrypto.BlockCryptKeyServerHalf) error {
    98  	bg.mtx.RLock()
    99  	defer bg.mtx.RUnlock()
   100  	source, ok := bg.blockMap[ptr]
   101  	if !ok {
   102  		return errors.New("Block doesn't exist in fake block map")
   103  	}
   104  	block.Set(source.block)
   105  	return nil
   106  }
   107  
   108  func (bg *fakeBlockGetter) assembleBlockLocal(ctx context.Context,
   109  	kmd libkey.KeyMetadata, ptr data.BlockPointer, block data.Block, buf []byte,
   110  	serverHalf kbfscrypto.BlockCryptKeyServerHalf) error {
   111  	return bg.assembleBlock(ctx, kmd, ptr, block, buf, serverHalf)
   112  }
   113  
   114  func TestBlockRetrievalWorkerBasic(t *testing.T) {
   115  	t.Log("Test the basic ability of a worker to return a block.")
   116  	bg := newFakeBlockGetter(false)
   117  	q := newBlockRetrievalQueue(
   118  		0, 1, 0, newTestBlockRetrievalConfig(t, bg, nil),
   119  		env.EmptyAppStateUpdater{})
   120  	require.NotNil(t, q)
   121  	defer endBlockRetrievalQueueTest(t, q)
   122  
   123  	ptr1 := makeRandomBlockPointer(t)
   124  	block1 := makeFakeFileBlock(t, false)
   125  	_, continueCh1 := bg.setBlockToReturn(ptr1, block1)
   126  
   127  	block := &data.FileBlock{}
   128  	ch := q.Request(
   129  		context.Background(), 1, makeKMD(), ptr1, block,
   130  		data.NoCacheEntry, BlockRequestSolo)
   131  	continueCh1 <- nil
   132  	err := <-ch
   133  	require.NoError(t, err)
   134  	require.Equal(t, block1, block)
   135  }
   136  
   137  func TestBlockRetrievalWorkerBasicSoloCached(t *testing.T) {
   138  	t.Log("Test the worker fetching and caching a solo block.")
   139  	bg := newFakeBlockGetter(false)
   140  	q := newBlockRetrievalQueue(
   141  		0, 1, 0, newTestBlockRetrievalConfig(t, bg, nil),
   142  		env.EmptyAppStateUpdater{})
   143  	require.NotNil(t, q)
   144  	defer endBlockRetrievalQueueTest(t, q)
   145  
   146  	ptr1 := makeRandomBlockPointer(t)
   147  	block1 := makeFakeFileBlock(t, false)
   148  	_, continueCh1 := bg.setBlockToReturn(ptr1, block1)
   149  
   150  	block := &data.FileBlock{}
   151  	ch := q.Request(
   152  		context.Background(), 1, makeKMD(), ptr1, block, data.TransientEntry,
   153  		BlockRequestSolo)
   154  	continueCh1 <- nil
   155  	err := <-ch
   156  	require.NoError(t, err)
   157  
   158  	_, err = q.config.BlockCache().Get(ptr1)
   159  	require.NoError(t, err)
   160  }
   161  
   162  func TestBlockRetrievalWorkerMultipleWorkers(t *testing.T) {
   163  	t.Log("Test the ability of multiple workers to retrieve concurrently.")
   164  	bg := newFakeBlockGetter(false)
   165  	q := newBlockRetrievalQueue(
   166  		2, 0, 0, newTestBlockRetrievalConfig(t, bg, nil),
   167  		env.EmptyAppStateUpdater{})
   168  	require.NotNil(t, q)
   169  	defer endBlockRetrievalQueueTest(t, q)
   170  
   171  	ptr1, ptr2 := makeRandomBlockPointer(t), makeRandomBlockPointer(t)
   172  	block1, block2 := makeFakeFileBlock(t, false), makeFakeFileBlock(t, false)
   173  	_, continueCh1 := bg.setBlockToReturn(ptr1, block1)
   174  	_, continueCh2 := bg.setBlockToReturn(ptr2, block2)
   175  
   176  	t.Log("Make 2 requests for 2 different blocks")
   177  	block := &data.FileBlock{}
   178  	// Set the base priority to be above the default on-demand
   179  	// fetching, so that the pre-prefetch request for a block doesn't
   180  	// override the other blocks' requests.
   181  	basePriority := defaultOnDemandRequestPriority + 1
   182  	req1Ch := q.Request(
   183  		context.Background(), basePriority, makeKMD(), ptr1, block,
   184  		data.NoCacheEntry, BlockRequestSolo)
   185  	req2Ch := q.Request(
   186  		context.Background(), basePriority, makeKMD(), ptr2, block,
   187  		data.NoCacheEntry, BlockRequestSolo)
   188  
   189  	t.Log("Allow the second request to complete before the first")
   190  	continueCh2 <- nil
   191  	err := <-req2Ch
   192  	require.NoError(t, err)
   193  	require.Equal(t, block2, block)
   194  
   195  	t.Log("Make another request for ptr2")
   196  	req2Ch = q.Request(
   197  		context.Background(), basePriority, makeKMD(), ptr2, block,
   198  		data.NoCacheEntry, BlockRequestSolo)
   199  	continueCh2 <- nil
   200  	err = <-req2Ch
   201  	require.NoError(t, err)
   202  	require.Equal(t, block2, block)
   203  
   204  	t.Log("Complete the ptr1 request")
   205  	continueCh1 <- nil
   206  	err = <-req1Ch
   207  	require.NoError(t, err)
   208  	require.Equal(t, block1, block)
   209  }
   210  
   211  func TestBlockRetrievalWorkerWithQueue(t *testing.T) {
   212  	t.Log("Test the ability of a worker and queue to work correctly together.")
   213  	bg := newFakeBlockGetter(false)
   214  	q := newBlockRetrievalQueue(
   215  		1, 0, 0, newTestBlockRetrievalConfig(t, bg, nil),
   216  		env.EmptyAppStateUpdater{})
   217  	require.NotNil(t, q)
   218  	defer endBlockRetrievalQueueTest(t, q)
   219  
   220  	ptr1, ptr2, ptr3 := makeRandomBlockPointer(t), makeRandomBlockPointer(t),
   221  		makeRandomBlockPointer(t)
   222  	block1, block2, block3 := makeFakeFileBlock(t, false),
   223  		makeFakeFileBlock(t, false), makeFakeFileBlock(t, false)
   224  	startCh1, continueCh1 := bg.setBlockToReturn(ptr1, block1)
   225  	_, continueCh2 := bg.setBlockToReturn(ptr2, block2)
   226  	_, continueCh3 := bg.setBlockToReturn(ptr3, block3)
   227  
   228  	t.Log("Make 3 retrievals for 3 different blocks. All retrievals after " +
   229  		"the first should be queued.")
   230  	block := &data.FileBlock{}
   231  	testBlock1 := &data.FileBlock{}
   232  	testBlock2 := &data.FileBlock{}
   233  	// Set the base priority to be above the default on-demand
   234  	// fetching, so that the pre-prefetch request for a block doesn't
   235  	// override the other blocks' requests.
   236  	basePriority := defaultOnDemandRequestPriority + 1
   237  	req1Ch := q.Request(
   238  		context.Background(), basePriority, makeKMD(), ptr1,
   239  		block, data.NoCacheEntry, BlockRequestSolo)
   240  	req2Ch := q.Request(
   241  		context.Background(), basePriority, makeKMD(), ptr2,
   242  		block, data.NoCacheEntry, BlockRequestSolo)
   243  	req3Ch := q.Request(
   244  		context.Background(), basePriority, makeKMD(), ptr3, testBlock1,
   245  		data.NoCacheEntry, BlockRequestSolo)
   246  	// Ensure the worker picks up the first request
   247  	<-startCh1
   248  	t.Log("Make a high priority request for the third block, which should " +
   249  		"complete next.")
   250  	req4Ch := q.Request(
   251  		context.Background(), basePriority+1, makeKMD(), ptr3, testBlock2,
   252  		data.NoCacheEntry, BlockRequestSolo)
   253  
   254  	t.Log("Allow the ptr1 retrieval to complete.")
   255  	continueCh1 <- nil
   256  	err := <-req1Ch
   257  	require.NoError(t, err)
   258  	require.Equal(t, block1, block)
   259  
   260  	t.Log("Allow the ptr3 retrieval to complete. Both waiting requests " +
   261  		"should complete.")
   262  	continueCh3 <- nil
   263  	err1 := <-req3Ch
   264  	err2 := <-req4Ch
   265  	require.NoError(t, err1)
   266  	require.NoError(t, err2)
   267  	require.Equal(t, block3, testBlock1)
   268  	require.Equal(t, block3, testBlock2)
   269  
   270  	t.Log("Complete the ptr2 retrieval.")
   271  	continueCh2 <- nil
   272  	err = <-req2Ch
   273  	require.NoError(t, err)
   274  	require.Equal(t, block2, block)
   275  }
   276  
   277  func TestBlockRetrievalWorkerCancel(t *testing.T) {
   278  	t.Log("Test the ability of a worker to handle a request cancelation.")
   279  	bg := newFakeBlockGetter(true)
   280  	q := newBlockRetrievalQueue(
   281  		0, 1, 0, newTestBlockRetrievalConfig(t, bg, nil),
   282  		env.EmptyAppStateUpdater{})
   283  	require.NotNil(t, q)
   284  	defer endBlockRetrievalQueueTest(t, q)
   285  
   286  	ptr1 := makeRandomBlockPointer(t)
   287  	block1 := makeFakeFileBlock(t, false)
   288  	// Don't need continueCh here.
   289  	_, _ = bg.setBlockToReturn(ptr1, block1)
   290  
   291  	block := &data.FileBlock{}
   292  	ctx, cancel := context.WithCancel(context.Background())
   293  	cancel()
   294  	ch := q.Request(
   295  		ctx, 1, makeKMD(), ptr1, block, data.NoCacheEntry, BlockRequestSolo)
   296  	err := <-ch
   297  	require.EqualError(t, err, context.Canceled.Error())
   298  }
   299  
   300  func TestBlockRetrievalWorkerShutdown(t *testing.T) {
   301  	t.Log("Test that worker shutdown works.")
   302  	bg := newFakeBlockGetter(false)
   303  	q := newBlockRetrievalQueue(
   304  		1, 0, 0, newTestBlockRetrievalConfig(t, bg, nil),
   305  		env.EmptyAppStateUpdater{})
   306  	require.NotNil(t, q)
   307  	defer endBlockRetrievalQueueTest(t, q)
   308  
   309  	w := q.workers[0]
   310  	require.NotNil(t, w)
   311  
   312  	ptr1 := makeRandomBlockPointer(t)
   313  	block1 := makeFakeFileBlock(t, false)
   314  	_, continueCh := bg.setBlockToReturn(ptr1, block1)
   315  
   316  	w.Shutdown()
   317  	block := &data.FileBlock{}
   318  	ctx, cancel := context.WithCancel(context.Background())
   319  	// Ensure the context loop is stopped so the test doesn't leak goroutines
   320  	defer cancel()
   321  	ch := q.Request(
   322  		ctx, 1, makeKMD(), ptr1, block, data.NoCacheEntry, BlockRequestSolo)
   323  	shutdown := false
   324  	select {
   325  	case <-ch:
   326  		t.Fatal("Expected not to retrieve a result from the Request.")
   327  	case continueCh <- nil:
   328  		t.Fatal("Expected the block getter not to be receiving.")
   329  	default:
   330  		shutdown = true
   331  	}
   332  	require.True(t, shutdown)
   333  
   334  	// Ensure the test completes in a reasonable time.
   335  	timer := time.NewTimer(10 * time.Second)
   336  	doneCh := make(chan struct{})
   337  	go func() {
   338  		w.Shutdown()
   339  		close(doneCh)
   340  	}()
   341  	select {
   342  	case <-timer.C:
   343  		t.Fatal("Expected another Shutdown not to block.")
   344  	case <-doneCh:
   345  	}
   346  }
   347  
   348  func TestBlockRetrievalWorkerPrefetchedPriorityElevation(t *testing.T) {
   349  	t.Log("Test that we can escalate the priority of a request and it " +
   350  		"correctly switches workers.")
   351  	bg := newFakeBlockGetter(false)
   352  	q := newBlockRetrievalQueue(
   353  		1, 1, 0, newTestBlockRetrievalConfig(t, bg, nil),
   354  		env.EmptyAppStateUpdater{})
   355  	require.NotNil(t, q)
   356  	defer endBlockRetrievalQueueTest(t, q)
   357  
   358  	t.Log("Setup source blocks")
   359  	ptr1, ptr2 := makeRandomBlockPointer(t), makeRandomBlockPointer(t)
   360  	block1, block2 := makeFakeFileBlock(t, false), makeFakeFileBlock(t, false)
   361  	_, continueCh1 := bg.setBlockToReturn(ptr1, block1)
   362  	_, continueCh2 := bg.setBlockToReturn(ptr2, block2)
   363  
   364  	t.Log("Make a low-priority request. This will get to the worker.")
   365  	testBlock1 := &data.FileBlock{}
   366  	req1Ch := q.Request(
   367  		context.Background(), 1, makeKMD(), ptr1, testBlock1,
   368  		data.NoCacheEntry, BlockRequestSolo)
   369  
   370  	t.Log("Make another low-priority request. This will block.")
   371  	testBlock2 := &data.FileBlock{}
   372  	req2Ch := q.Request(
   373  		context.Background(), 1, makeKMD(), ptr2, testBlock2,
   374  		data.NoCacheEntry, BlockRequestSolo)
   375  
   376  	t.Log("Make an on-demand request for the same block as the blocked " +
   377  		"request.")
   378  	testBlock3 := &data.FileBlock{}
   379  	req3Ch := q.Request(
   380  		context.Background(), defaultOnDemandRequestPriority,
   381  		makeKMD(), ptr2, testBlock3, data.NoCacheEntry, BlockRequestSolo)
   382  
   383  	t.Log("Release the requests for the second block first. " +
   384  		"Since the prefetch worker is still blocked, this confirms that the " +
   385  		"escalation to an on-demand worker was successful.")
   386  	continueCh2 <- nil
   387  	err := <-req3Ch
   388  	require.NoError(t, err)
   389  	require.Equal(t, testBlock3, block2)
   390  	err = <-req2Ch
   391  	require.NoError(t, err)
   392  	require.Equal(t, testBlock2, block2)
   393  
   394  	t.Log("Allow the initial ptr1 request to complete.")
   395  	continueCh1 <- nil
   396  	err = <-req1Ch
   397  	require.NoError(t, err)
   398  	require.Equal(t, testBlock1, block1)
   399  }
   400  
   401  func TestBlockRetrievalWorkerStopIfFull(t *testing.T) {
   402  	ctx, cancel := context.WithTimeout(
   403  		context.Background(), individualTestTimeout)
   404  	defer cancel()
   405  	dbc, dbcConfig := initDiskBlockCacheTest(t)
   406  	defer dbc.Shutdown(ctx)
   407  
   408  	bg := newFakeBlockGetter(false)
   409  	q := newBlockRetrievalQueue(
   410  		1, 1, 0, newTestBlockRetrievalConfig(t, bg, dbc),
   411  		env.EmptyAppStateUpdater{})
   412  	require.NotNil(t, q)
   413  	<-q.TogglePrefetcher(false, nil, nil)
   414  	defer endBlockRetrievalQueueTest(t, q)
   415  
   416  	ptr := makeRandomBlockPointer(t)
   417  	syncCache := dbc.syncCache
   418  	workingCache := dbc.workingSetCache
   419  
   420  	t.Log("Set the cache maximum bytes to the current total.")
   421  	syncBytes, workingBytes := testGetDiskCacheBytes(syncCache, workingCache)
   422  	limiter := dbcConfig.DiskLimiter().(*backpressureDiskLimiter)
   423  	setLimiterLimits(limiter, syncBytes, workingBytes)
   424  
   425  	t.Log("Request with stop-if-full, when full")
   426  	testBlock := &data.FileBlock{}
   427  	req := q.Request(
   428  		ctx, 1, makeKMD(), ptr, testBlock, data.NoCacheEntry,
   429  		BlockRequestPrefetchUntilFull)
   430  	select {
   431  	case err := <-req:
   432  		require.IsType(t, DiskCacheTooFullForBlockError{}, err)
   433  	case <-ctx.Done():
   434  		require.FailNow(t, ctx.Err().Error())
   435  	}
   436  
   437  	t.Log("Request without stop-if-full, when full")
   438  	block := makeFakeFileBlock(t, false)
   439  	startCh, continueCh := bg.setBlockToReturn(ptr, block)
   440  	req = q.Request(
   441  		ctx, 1, makeKMD(), ptr, testBlock, data.NoCacheEntry,
   442  		BlockRequestSolo)
   443  	<-startCh
   444  	continueCh <- nil
   445  	select {
   446  	case err := <-req:
   447  		require.NoError(t, err)
   448  	case <-ctx.Done():
   449  		require.FailNow(t, ctx.Err().Error())
   450  	}
   451  }