github.com/m3db/m3@v1.5.0/src/aggregator/client/writer_test.go (about)

     1  // Copyright (c) 2018 Uber Technologies, Inc.
     2  //
     3  // Permission is hereby granted, free of charge, to any person obtaining a copy
     4  // of this software and associated documentation files (the "Software"), to deal
     5  // in the Software without restriction, including without limitation the rights
     6  // to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
     7  // copies of the Software, and to permit persons to whom the Software is
     8  // furnished to do so, subject to the following conditions:
     9  //
    10  // The above copyright notice and this permission notice shall be included in
    11  // all copies or substantial portions of the Software.
    12  //
    13  // THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
    14  // IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
    15  // FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
    16  // AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
    17  // LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
    18  // OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN
    19  // THE SOFTWARE.
    20  
    21  package client
    22  
    23  import (
    24  	"bytes"
    25  	"errors"
    26  	"fmt"
    27  	"io"
    28  	"math"
    29  	"sort"
    30  	"strings"
    31  	"sync"
    32  	"testing"
    33  	"time"
    34  
    35  	"github.com/m3db/m3/src/metrics/encoding"
    36  	"github.com/m3db/m3/src/metrics/encoding/protobuf"
    37  	"github.com/m3db/m3/src/metrics/metadata"
    38  	"github.com/m3db/m3/src/metrics/metric"
    39  	"github.com/m3db/m3/src/metrics/metric/aggregated"
    40  	"github.com/m3db/m3/src/metrics/metric/id"
    41  	"github.com/m3db/m3/src/metrics/metric/unaggregated"
    42  	"github.com/m3db/m3/src/x/clock"
    43  	"github.com/m3db/m3/src/x/instrument"
    44  
    45  	"github.com/golang/mock/gomock"
    46  	"github.com/stretchr/testify/require"
    47  	"github.com/uber-go/tally"
    48  )
    49  
    50  func TestWriterWriteClosed(t *testing.T) {
    51  	payload := payloadUnion{
    52  		payloadType: untimedType,
    53  		untimed: untimedPayload{
    54  			metric:    testCounter,
    55  			metadatas: testStagedMetadatas,
    56  		},
    57  	}
    58  	w := newInstanceWriter(testPlacementInstance, testOptions()).(*writer)
    59  	w.closed = true
    60  	require.Equal(t, errInstanceWriterClosed, w.Write(0, payload))
    61  }
    62  
    63  func TestWriterWriteUntimedCounterEncodeError(t *testing.T) {
    64  	ctrl := gomock.NewController(t)
    65  	defer ctrl.Finish()
    66  
    67  	errTestEncodeMetric := errors.New("error encoding metrics")
    68  	w := newInstanceWriter(testPlacementInstance, testOptions()).(*writer)
    69  	w.newLockedEncoderFn = func(protobuf.UnaggregatedOptions) *lockedEncoder {
    70  		encoder := protobuf.NewMockUnaggregatedEncoder(ctrl)
    71  		encoder.EXPECT().Len().Return(0)
    72  		encoder.EXPECT().EncodeMessage(encoding.UnaggregatedMessageUnion{
    73  			Type: encoding.CounterWithMetadatasType,
    74  			CounterWithMetadatas: unaggregated.CounterWithMetadatas{
    75  				Counter:         testCounter.Counter(),
    76  				StagedMetadatas: testStagedMetadatas,
    77  			},
    78  		}).Return(errTestEncodeMetric)
    79  		encoder.EXPECT().Truncate(0).Return(nil)
    80  		return &lockedEncoder{UnaggregatedEncoder: encoder}
    81  	}
    82  
    83  	payload := payloadUnion{
    84  		payloadType: untimedType,
    85  		untimed: untimedPayload{
    86  			metric:    testCounter,
    87  			metadatas: testStagedMetadatas,
    88  		},
    89  	}
    90  	require.Equal(t, errTestEncodeMetric, w.Write(0, payload))
    91  }
    92  
    93  func TestWriterWriteUntimedCounterEncoderExists(t *testing.T) {
    94  	ctrl := gomock.NewController(t)
    95  	defer ctrl.Finish()
    96  
    97  	w := newInstanceWriter(testPlacementInstance, testOptions()).(*writer)
    98  	encoder := protobuf.NewMockUnaggregatedEncoder(ctrl)
    99  	gomock.InOrder(
   100  		encoder.EXPECT().Len().Return(0),
   101  		encoder.EXPECT().EncodeMessage(encoding.UnaggregatedMessageUnion{
   102  			Type: encoding.CounterWithMetadatasType,
   103  			CounterWithMetadatas: unaggregated.CounterWithMetadatas{
   104  				Counter:         testCounter.Counter(),
   105  				StagedMetadatas: testStagedMetadatas,
   106  			},
   107  		}).Return(nil),
   108  		encoder.EXPECT().Len().Return(4),
   109  	)
   110  	w.encodersByShard[0] = &lockedEncoder{UnaggregatedEncoder: encoder}
   111  
   112  	payload := payloadUnion{
   113  		payloadType: untimedType,
   114  		untimed: untimedPayload{
   115  			metric:    testCounter,
   116  			metadatas: testStagedMetadatas,
   117  		},
   118  	}
   119  	require.NoError(t, w.Write(0, payload))
   120  	require.Equal(t, 1, len(w.encodersByShard))
   121  }
   122  
   123  func TestWriterWriteUntimedCounterEncoderDoesNotExist(t *testing.T) {
   124  	ctrl := gomock.NewController(t)
   125  	defer ctrl.Finish()
   126  
   127  	encoder := protobuf.NewMockUnaggregatedEncoder(ctrl)
   128  	gomock.InOrder(
   129  		encoder.EXPECT().Len().Return(3),
   130  		encoder.EXPECT().EncodeMessage(encoding.UnaggregatedMessageUnion{
   131  			Type: encoding.CounterWithMetadatasType,
   132  			CounterWithMetadatas: unaggregated.CounterWithMetadatas{
   133  				Counter:         testCounter.Counter(),
   134  				StagedMetadatas: testStagedMetadatas,
   135  			},
   136  		}).Return(nil),
   137  		encoder.EXPECT().Len().Return(7),
   138  	)
   139  	w := newInstanceWriter(testPlacementInstance, testOptions()).(*writer)
   140  	w.newLockedEncoderFn = func(protobuf.UnaggregatedOptions) *lockedEncoder {
   141  		return &lockedEncoder{UnaggregatedEncoder: encoder}
   142  	}
   143  
   144  	payload := payloadUnion{
   145  		payloadType: untimedType,
   146  		untimed: untimedPayload{
   147  			metric:    testCounter,
   148  			metadatas: testStagedMetadatas,
   149  		},
   150  	}
   151  	require.NoError(t, w.Write(0, payload))
   152  }
   153  
   154  func TestWriterWriteUntimedCounterWithFlushingZeroSizeBefore(t *testing.T) {
   155  	ctrl := gomock.NewController(t)
   156  	defer ctrl.Finish()
   157  
   158  	var (
   159  		stream      = protobuf.NewBuffer([]byte{1, 2, 3, 4, 5, 6, 7}, nil)
   160  		enqueuedBuf protobuf.Buffer
   161  	)
   162  	encoder := protobuf.NewMockUnaggregatedEncoder(ctrl)
   163  	gomock.InOrder(
   164  		encoder.EXPECT().Len().Return(0),
   165  		encoder.EXPECT().EncodeMessage(encoding.UnaggregatedMessageUnion{
   166  			Type: encoding.CounterWithMetadatasType,
   167  			CounterWithMetadatas: unaggregated.CounterWithMetadatas{
   168  				Counter:         testCounter.Counter(),
   169  				StagedMetadatas: testStagedMetadatas,
   170  			},
   171  		}).Return(nil),
   172  		encoder.EXPECT().Len().Return(7),
   173  		encoder.EXPECT().Relinquish().Return(stream),
   174  	)
   175  	queue := NewMockinstanceQueue(ctrl)
   176  	queue.EXPECT().
   177  		Enqueue(gomock.Any()).
   178  		DoAndReturn(func(buf protobuf.Buffer) error {
   179  			enqueuedBuf = buf
   180  			return nil
   181  		})
   182  	w := newInstanceWriter(testPlacementInstance, testOptions().SetMaxBatchSize(3)).(*writer)
   183  	w.queue = queue
   184  	w.newLockedEncoderFn = func(protobuf.UnaggregatedOptions) *lockedEncoder {
   185  		return &lockedEncoder{UnaggregatedEncoder: encoder}
   186  	}
   187  
   188  	payload := payloadUnion{
   189  		payloadType: untimedType,
   190  		untimed: untimedPayload{
   191  			metric:    testCounter,
   192  			metadatas: testStagedMetadatas,
   193  		},
   194  	}
   195  	require.NoError(t, w.Write(0, payload))
   196  
   197  	enc, exists := w.encodersByShard[0]
   198  	require.True(t, exists)
   199  	require.NotNil(t, enc)
   200  	require.Equal(t, 1, len(w.encodersByShard))
   201  	require.Equal(t, []byte{1, 2, 3, 4, 5, 6, 7}, enqueuedBuf.Bytes())
   202  }
   203  
   204  func TestWriterWriteUntimedCounterWithFlushingPositiveSizeBefore(t *testing.T) {
   205  	ctrl := gomock.NewController(t)
   206  	defer ctrl.Finish()
   207  
   208  	var (
   209  		stream      = protobuf.NewBuffer([]byte{1, 2, 3, 4, 5, 6, 7}, nil)
   210  		enqueuedBuf protobuf.Buffer
   211  	)
   212  	encoder := protobuf.NewMockUnaggregatedEncoder(ctrl)
   213  	gomock.InOrder(
   214  		encoder.EXPECT().Len().Return(3),
   215  		encoder.EXPECT().EncodeMessage(encoding.UnaggregatedMessageUnion{
   216  			Type: encoding.CounterWithMetadatasType,
   217  			CounterWithMetadatas: unaggregated.CounterWithMetadatas{
   218  				Counter:         testCounter.Counter(),
   219  				StagedMetadatas: testStagedMetadatas,
   220  			},
   221  		}).Return(nil),
   222  		encoder.EXPECT().Len().Return(7),
   223  		encoder.EXPECT().Relinquish().Return(stream),
   224  	)
   225  	queue := NewMockinstanceQueue(ctrl)
   226  	queue.EXPECT().
   227  		Enqueue(gomock.Any()).
   228  		DoAndReturn(func(buf protobuf.Buffer) error {
   229  			enqueuedBuf = buf
   230  			return nil
   231  		})
   232  	w := newInstanceWriter(testPlacementInstance, testOptions().SetMaxBatchSize(3)).(*writer)
   233  	w.queue = queue
   234  	w.newLockedEncoderFn = func(protobuf.UnaggregatedOptions) *lockedEncoder {
   235  		return &lockedEncoder{UnaggregatedEncoder: encoder}
   236  	}
   237  
   238  	payload := payloadUnion{
   239  		payloadType: untimedType,
   240  		untimed: untimedPayload{
   241  			metric:    testCounter,
   242  			metadatas: testStagedMetadatas,
   243  		},
   244  	}
   245  	require.NoError(t, w.Write(0, payload))
   246  
   247  	enc, exists := w.encodersByShard[0]
   248  	require.True(t, exists)
   249  	require.NotNil(t, enc)
   250  	require.Equal(t, 1, len(w.encodersByShard))
   251  	require.Equal(t, []byte{0x1, 0x2, 0x3, 0x4, 0x5, 0x6, 0x7}, enqueuedBuf.Bytes())
   252  }
   253  
   254  func TestWriterWriteUntimedBatchTimerNoBatchSizeLimit(t *testing.T) {
   255  	ctrl := gomock.NewController(t)
   256  	defer ctrl.Finish()
   257  
   258  	numValues := 65536
   259  	timerValues := make([]float64, numValues)
   260  	for i := 0; i < numValues; i++ {
   261  		timerValues[i] = float64(i)
   262  	}
   263  	testLargeBatchTimer := unaggregated.MetricUnion{
   264  		Type:          metric.TimerType,
   265  		ID:            []byte("testLargeBatchTimer"),
   266  		BatchTimerVal: timerValues,
   267  	}
   268  	encoder := protobuf.NewMockUnaggregatedEncoder(ctrl)
   269  	gomock.InOrder(
   270  		encoder.EXPECT().Len().Return(3),
   271  		encoder.EXPECT().EncodeMessage(encoding.UnaggregatedMessageUnion{
   272  			Type: encoding.BatchTimerWithMetadatasType,
   273  			BatchTimerWithMetadatas: unaggregated.BatchTimerWithMetadatas{
   274  				BatchTimer:      testLargeBatchTimer.BatchTimer(),
   275  				StagedMetadatas: testStagedMetadatas,
   276  			},
   277  		}).Return(nil),
   278  		encoder.EXPECT().Len().Return(7),
   279  	)
   280  	opts := testOptions().SetMaxTimerBatchSize(0)
   281  	w := newInstanceWriter(testPlacementInstance, opts).(*writer)
   282  	w.newLockedEncoderFn = func(protobuf.UnaggregatedOptions) *lockedEncoder {
   283  		return &lockedEncoder{UnaggregatedEncoder: encoder}
   284  	}
   285  
   286  	payload := payloadUnion{
   287  		payloadType: untimedType,
   288  		untimed: untimedPayload{
   289  			metric:    testLargeBatchTimer,
   290  			metadatas: testStagedMetadatas,
   291  		},
   292  	}
   293  	require.NoError(t, w.Write(0, payload))
   294  }
   295  
   296  func TestWriterWriteUntimedBatchTimerSmallBatchSize(t *testing.T) {
   297  	ctrl := gomock.NewController(t)
   298  	defer ctrl.Finish()
   299  
   300  	encoder := protobuf.NewMockUnaggregatedEncoder(ctrl)
   301  	gomock.InOrder(
   302  		encoder.EXPECT().Len().Return(3),
   303  		encoder.EXPECT().EncodeMessage(encoding.UnaggregatedMessageUnion{
   304  			Type: encoding.BatchTimerWithMetadatasType,
   305  			BatchTimerWithMetadatas: unaggregated.BatchTimerWithMetadatas{
   306  				BatchTimer:      testBatchTimer.BatchTimer(),
   307  				StagedMetadatas: testStagedMetadatas,
   308  			},
   309  		}).Return(nil),
   310  		encoder.EXPECT().Len().Return(7),
   311  	)
   312  	opts := testOptions().SetMaxTimerBatchSize(140)
   313  	w := newInstanceWriter(testPlacementInstance, opts).(*writer)
   314  	w.newLockedEncoderFn = func(protobuf.UnaggregatedOptions) *lockedEncoder {
   315  		return &lockedEncoder{UnaggregatedEncoder: encoder}
   316  	}
   317  
   318  	payload := payloadUnion{
   319  		payloadType: untimedType,
   320  		untimed: untimedPayload{
   321  			metric:    testBatchTimer,
   322  			metadatas: testStagedMetadatas,
   323  		},
   324  	}
   325  	require.NoError(t, w.Write(0, payload))
   326  }
   327  
   328  func TestWriterWriteUntimedBatchTimerLargeBatchSize(t *testing.T) {
   329  	ctrl := gomock.NewController(t)
   330  	defer ctrl.Finish()
   331  
   332  	numValues := 65536
   333  	timerValues := make([]float64, numValues)
   334  	for i := 0; i < numValues; i++ {
   335  		timerValues[i] = float64(i)
   336  	}
   337  	testLargeBatchTimer := unaggregated.MetricUnion{
   338  		Type:          metric.TimerType,
   339  		ID:            []byte("testLargeBatchTimer"),
   340  		BatchTimerVal: timerValues,
   341  	}
   342  
   343  	var (
   344  		msgTypeRes         []encoding.UnaggregatedMessageType
   345  		idRes              []id.RawID
   346  		valueRes           [][]float64
   347  		metadataRes        []metadata.StagedMetadatas
   348  		maxBatchSize       = 140
   349  		expectedNumBatches = int(math.Ceil(float64(numValues) / float64(maxBatchSize)))
   350  	)
   351  	encoder := protobuf.NewMockUnaggregatedEncoder(ctrl)
   352  	encoder.EXPECT().Len().Return(7).MinTimes(2)
   353  	encoder.EXPECT().
   354  		EncodeMessage(gomock.Any()).
   355  		DoAndReturn(func(msg encoding.UnaggregatedMessageUnion) error {
   356  			msgTypeRes = append(msgTypeRes, msg.Type)
   357  			idRes = append(idRes, msg.BatchTimerWithMetadatas.ID)
   358  			valueRes = append(valueRes, msg.BatchTimerWithMetadatas.Values)
   359  			metadataRes = append(metadataRes, msg.BatchTimerWithMetadatas.StagedMetadatas)
   360  			return nil
   361  		}).Times(expectedNumBatches)
   362  	encoder.EXPECT().Relinquish()
   363  
   364  	opts := testOptions().SetMaxTimerBatchSize(maxBatchSize)
   365  	w := newInstanceWriter(testPlacementInstance, opts).(*writer)
   366  	w.newLockedEncoderFn = func(protobuf.UnaggregatedOptions) *lockedEncoder {
   367  		return &lockedEncoder{UnaggregatedEncoder: encoder}
   368  	}
   369  
   370  	payload := payloadUnion{
   371  		payloadType: untimedType,
   372  		untimed: untimedPayload{
   373  			metric:    testLargeBatchTimer,
   374  			metadatas: testStagedMetadatas,
   375  		},
   376  	}
   377  	require.NoError(t, w.Write(0, payload))
   378  	require.NoError(t, w.Flush())
   379  
   380  	var (
   381  		expectedMsgTypes  []encoding.UnaggregatedMessageType
   382  		expectedIDs       []id.RawID
   383  		expectedValues    [][]float64
   384  		expectedMetadatas []metadata.StagedMetadatas
   385  	)
   386  	for i := 0; i < expectedNumBatches; i++ {
   387  		start := i * maxBatchSize
   388  		end := start + maxBatchSize
   389  		if end > numValues {
   390  			end = numValues
   391  		}
   392  		expectedMsgTypes = append(expectedMsgTypes, encoding.BatchTimerWithMetadatasType)
   393  		expectedValues = append(expectedValues, timerValues[start:end])
   394  		expectedIDs = append(expectedIDs, id.RawID("testLargeBatchTimer"))
   395  		expectedMetadatas = append(expectedMetadatas, testStagedMetadatas)
   396  	}
   397  	require.Equal(t, expectedMsgTypes, msgTypeRes)
   398  	require.Equal(t, expectedIDs, idRes)
   399  	require.Equal(t, expectedValues, valueRes)
   400  	require.Equal(t, expectedMetadatas, metadataRes)
   401  }
   402  
   403  func TestWriterWriteUntimedLargeBatchTimerUsesMultipleBuffers(t *testing.T) {
   404  	const (
   405  		numValues  = 1400
   406  		testIDName = "testLargeBatchTimer"
   407  	)
   408  
   409  	timerValues := make([]float64, numValues)
   410  	for i := 0; i < numValues; i++ {
   411  		timerValues[i] = float64(i)
   412  	}
   413  
   414  	var (
   415  		testLargeBatchTimer = unaggregated.MetricUnion{
   416  			Type:          metric.TimerType,
   417  			ID:            []byte(testIDName),
   418  			BatchTimerVal: timerValues,
   419  		}
   420  		payload = payloadUnion{
   421  			payloadType: untimedType,
   422  			untimed: untimedPayload{
   423  				metric:    testLargeBatchTimer,
   424  				metadatas: testStagedMetadatas,
   425  			},
   426  		}
   427  		testScope = tally.NewTestScope("", nil)
   428  		iOpts     = instrument.NewOptions().SetMetricsScope(testScope)
   429  		opts      = testOptions().
   430  				SetMaxBatchSize(1000).
   431  				SetMaxTimerBatchSize(10).
   432  				SetInstrumentOptions(iOpts)
   433  
   434  		w            = newInstanceWriter(testPlacementInstance, opts).(*writer)
   435  		q            = w.queue.(*queue)
   436  		payloadCount int
   437  	)
   438  
   439  	q.writeFn = func(payload []byte) error {
   440  		payloadCount += strings.Count(string(payload), testIDName)
   441  		return nil
   442  	}
   443  
   444  	require.NoError(t, w.Write(0, payload))
   445  	require.NoError(t, w.Flush())
   446  	time.Sleep(1 * time.Second) // TODO: remove once queue is sync
   447  	require.NoError(t, w.Close())
   448  
   449  	enqueuedCounter := testScope.Snapshot().Counters()["buffers+action=enqueued"]
   450  	require.NotNil(t, enqueuedCounter)
   451  	// Expect 1 byte buffer to be enqueued to write to network,
   452  	// but timer itself should be split to multiple protobuf payloads.
   453  	require.Equal(t, int64(1), enqueuedCounter.Value())
   454  	require.Equal(t, 140, payloadCount)
   455  }
   456  
   457  func TestWriterWriteUntimedBatchTimerWriteError(t *testing.T) {
   458  	ctrl := gomock.NewController(t)
   459  	defer ctrl.Finish()
   460  
   461  	numValues := 7
   462  	timerValues := make([]float64, numValues)
   463  	for i := 0; i < numValues; i++ {
   464  		timerValues[i] = float64(i)
   465  	}
   466  	testLargeBatchTimer := unaggregated.MetricUnion{
   467  		Type:          metric.TimerType,
   468  		ID:            []byte("testLargeBatchTimer"),
   469  		BatchTimerVal: timerValues,
   470  	}
   471  
   472  	errTestWrite = errors.New("test write error")
   473  	encoder := protobuf.NewMockUnaggregatedEncoder(ctrl)
   474  	gomock.InOrder(
   475  		encoder.EXPECT().Len().Return(3),
   476  		encoder.EXPECT().
   477  			EncodeMessage(gomock.Any()).
   478  			Return(nil),
   479  		encoder.EXPECT().
   480  			EncodeMessage(gomock.Any()).
   481  			Return(errTestWrite),
   482  
   483  		encoder.EXPECT().Truncate(3).Return(nil),
   484  	)
   485  	opts := testOptions().SetMaxTimerBatchSize(3)
   486  	w := newInstanceWriter(testPlacementInstance, opts).(*writer)
   487  	w.newLockedEncoderFn = func(protobuf.UnaggregatedOptions) *lockedEncoder {
   488  		return &lockedEncoder{UnaggregatedEncoder: encoder}
   489  	}
   490  
   491  	payload := payloadUnion{
   492  		payloadType: untimedType,
   493  		untimed: untimedPayload{
   494  			metric:    testLargeBatchTimer,
   495  			metadatas: testStagedMetadatas,
   496  		},
   497  	}
   498  	require.Equal(t, errTestWrite, w.Write(0, payload))
   499  }
   500  
   501  func TestWriterWriteUntimedBatchTimerEnqueueError(t *testing.T) {
   502  	ctrl := gomock.NewController(t)
   503  	defer ctrl.Finish()
   504  
   505  	errTestEnqueue := errors.New("test enqueue error")
   506  	queue := NewMockinstanceQueue(ctrl)
   507  	queue.EXPECT().Enqueue(gomock.Any()).Return(errTestEnqueue)
   508  	opts := testOptions().
   509  		SetMaxTimerBatchSize(1).
   510  		SetMaxBatchSize(1)
   511  	w := newInstanceWriter(testPlacementInstance, opts).(*writer)
   512  	w.queue = queue
   513  
   514  	payload := payloadUnion{
   515  		payloadType: untimedType,
   516  		untimed: untimedPayload{
   517  			metric:    testBatchTimer,
   518  			metadatas: testStagedMetadatas,
   519  		},
   520  	}
   521  	require.Equal(t, errTestEnqueue, w.Write(0, payload))
   522  }
   523  
   524  func TestWriterWriteUntimedGauge(t *testing.T) {
   525  	ctrl := gomock.NewController(t)
   526  	defer ctrl.Finish()
   527  
   528  	encoder := protobuf.NewMockUnaggregatedEncoder(ctrl)
   529  	gomock.InOrder(
   530  		encoder.EXPECT().Len().Return(3),
   531  		encoder.EXPECT().EncodeMessage(encoding.UnaggregatedMessageUnion{
   532  			Type: encoding.GaugeWithMetadatasType,
   533  			GaugeWithMetadatas: unaggregated.GaugeWithMetadatas{
   534  				Gauge:           testGauge.Gauge(),
   535  				StagedMetadatas: testStagedMetadatas,
   536  			},
   537  		}).Return(nil),
   538  		encoder.EXPECT().Len().Return(7),
   539  	)
   540  	w := newInstanceWriter(testPlacementInstance, testOptions()).(*writer)
   541  	w.newLockedEncoderFn = func(protobuf.UnaggregatedOptions) *lockedEncoder {
   542  		return &lockedEncoder{UnaggregatedEncoder: encoder}
   543  	}
   544  
   545  	payload := payloadUnion{
   546  		payloadType: untimedType,
   547  		untimed: untimedPayload{
   548  			metric:    testGauge,
   549  			metadatas: testStagedMetadatas,
   550  		},
   551  	}
   552  	require.NoError(t, w.Write(0, payload))
   553  }
   554  
   555  func TestWriterWriteForwardedWithFlushingZeroSizeBefore(t *testing.T) {
   556  	ctrl := gomock.NewController(t)
   557  	defer ctrl.Finish()
   558  
   559  	var (
   560  		stream      = protobuf.NewBuffer([]byte{1, 2, 3, 4, 5, 6, 7}, nil)
   561  		enqueuedBuf protobuf.Buffer
   562  	)
   563  	encoder := protobuf.NewMockUnaggregatedEncoder(ctrl)
   564  	gomock.InOrder(
   565  		encoder.EXPECT().Len().Return(0),
   566  		encoder.EXPECT().EncodeMessage(encoding.UnaggregatedMessageUnion{
   567  			Type: encoding.ForwardedMetricWithMetadataType,
   568  			ForwardedMetricWithMetadata: aggregated.ForwardedMetricWithMetadata{
   569  				ForwardedMetric: testForwarded,
   570  				ForwardMetadata: testForwardMetadata,
   571  			},
   572  		}).Return(nil),
   573  		encoder.EXPECT().Len().Return(7),
   574  		encoder.EXPECT().Relinquish().Return(stream),
   575  	)
   576  	queue := NewMockinstanceQueue(ctrl)
   577  	queue.EXPECT().
   578  		Enqueue(gomock.Any()).
   579  		DoAndReturn(func(buf protobuf.Buffer) error {
   580  			enqueuedBuf = buf
   581  			return nil
   582  		})
   583  	w := newInstanceWriter(testPlacementInstance, testOptions().SetMaxBatchSize(3)).(*writer)
   584  	w.queue = queue
   585  	w.newLockedEncoderFn = func(protobuf.UnaggregatedOptions) *lockedEncoder {
   586  		return &lockedEncoder{UnaggregatedEncoder: encoder}
   587  	}
   588  
   589  	payload := payloadUnion{
   590  		payloadType: forwardedType,
   591  		forwarded: forwardedPayload{
   592  			metric:   testForwarded,
   593  			metadata: testForwardMetadata,
   594  		},
   595  	}
   596  	require.NoError(t, w.Write(0, payload))
   597  
   598  	enc, exists := w.encodersByShard[0]
   599  	require.True(t, exists)
   600  	require.NotNil(t, enc)
   601  	require.Equal(t, 1, len(w.encodersByShard))
   602  	require.Equal(t, []byte{1, 2, 3, 4, 5, 6, 7}, enqueuedBuf.Bytes())
   603  }
   604  
   605  func TestWriterWriteForwardedWithFlushingPositiveSizeBefore(t *testing.T) {
   606  	ctrl := gomock.NewController(t)
   607  	defer ctrl.Finish()
   608  
   609  	var (
   610  		stream      = protobuf.NewBuffer([]byte{1, 2, 3, 4, 5, 6, 7}, nil)
   611  		enqueuedBuf protobuf.Buffer
   612  	)
   613  	encoder := protobuf.NewMockUnaggregatedEncoder(ctrl)
   614  	gomock.InOrder(
   615  		encoder.EXPECT().Len().Return(3),
   616  		encoder.EXPECT().EncodeMessage(encoding.UnaggregatedMessageUnion{
   617  			Type: encoding.ForwardedMetricWithMetadataType,
   618  			ForwardedMetricWithMetadata: aggregated.ForwardedMetricWithMetadata{
   619  				ForwardedMetric: testForwarded,
   620  				ForwardMetadata: testForwardMetadata,
   621  			},
   622  		}).Return(nil),
   623  		encoder.EXPECT().Len().Return(7),
   624  		encoder.EXPECT().Relinquish().Return(stream),
   625  	)
   626  	queue := NewMockinstanceQueue(ctrl)
   627  	queue.EXPECT().
   628  		Enqueue(gomock.Any()).
   629  		DoAndReturn(func(buf protobuf.Buffer) error {
   630  			enqueuedBuf = buf
   631  			return nil
   632  		})
   633  	w := newInstanceWriter(testPlacementInstance, testOptions().SetMaxBatchSize(3)).(*writer)
   634  	w.queue = queue
   635  	w.newLockedEncoderFn = func(protobuf.UnaggregatedOptions) *lockedEncoder {
   636  		return &lockedEncoder{UnaggregatedEncoder: encoder}
   637  	}
   638  
   639  	payload := payloadUnion{
   640  		payloadType: forwardedType,
   641  		forwarded: forwardedPayload{
   642  			metric:   testForwarded,
   643  			metadata: testForwardMetadata,
   644  		},
   645  	}
   646  	require.NoError(t, w.Write(0, payload))
   647  
   648  	enc, exists := w.encodersByShard[0]
   649  	require.True(t, exists)
   650  	require.NotNil(t, enc)
   651  	require.Equal(t, 1, len(w.encodersByShard))
   652  	require.Equal(t, []byte{0x1, 0x2, 0x3, 0x4, 0x5, 0x6, 0x7}, enqueuedBuf.Bytes())
   653  }
   654  
   655  func TestWriterWriteForwardedEncodeError(t *testing.T) {
   656  	ctrl := gomock.NewController(t)
   657  	defer ctrl.Finish()
   658  
   659  	errTestEncodeMetric := errors.New("error encoding metrics")
   660  	w := newInstanceWriter(testPlacementInstance, testOptions()).(*writer)
   661  	w.newLockedEncoderFn = func(protobuf.UnaggregatedOptions) *lockedEncoder {
   662  		encoder := protobuf.NewMockUnaggregatedEncoder(ctrl)
   663  		encoder.EXPECT().Len().Return(0)
   664  		encoder.EXPECT().EncodeMessage(encoding.UnaggregatedMessageUnion{
   665  			Type: encoding.ForwardedMetricWithMetadataType,
   666  			ForwardedMetricWithMetadata: aggregated.ForwardedMetricWithMetadata{
   667  				ForwardedMetric: testForwarded,
   668  				ForwardMetadata: testForwardMetadata,
   669  			},
   670  		}).Return(errTestEncodeMetric)
   671  		encoder.EXPECT().Truncate(0).Return(nil)
   672  		return &lockedEncoder{UnaggregatedEncoder: encoder}
   673  	}
   674  
   675  	payload := payloadUnion{
   676  		payloadType: forwardedType,
   677  		forwarded: forwardedPayload{
   678  			metric:   testForwarded,
   679  			metadata: testForwardMetadata,
   680  		},
   681  	}
   682  	require.Equal(t, errTestEncodeMetric, w.Write(0, payload))
   683  }
   684  
   685  func TestWriterWriteForwardedEnqueueError(t *testing.T) {
   686  	ctrl := gomock.NewController(t)
   687  	defer ctrl.Finish()
   688  
   689  	errTestEnqueue := errors.New("test enqueue error")
   690  	queue := NewMockinstanceQueue(ctrl)
   691  	queue.EXPECT().Enqueue(gomock.Any()).Return(errTestEnqueue)
   692  	opts := testOptions().
   693  		SetMaxTimerBatchSize(1).
   694  		SetMaxBatchSize(1)
   695  	w := newInstanceWriter(testPlacementInstance, opts).(*writer)
   696  	w.queue = queue
   697  
   698  	payload := payloadUnion{
   699  		payloadType: forwardedType,
   700  		forwarded: forwardedPayload{
   701  			metric:   testForwarded,
   702  			metadata: testForwardMetadata,
   703  		},
   704  	}
   705  	require.Equal(t, errTestEnqueue, w.Write(0, payload))
   706  }
   707  
   708  func TestWriterFlushClosed(t *testing.T) {
   709  	w := newInstanceWriter(testPlacementInstance, testOptions()).(*writer)
   710  	w.closed = true
   711  	require.Equal(t, errInstanceWriterClosed, w.Flush())
   712  }
   713  
   714  func TestWriterFlushPartialError(t *testing.T) {
   715  	ctrl := gomock.NewController(t)
   716  	defer ctrl.Finish()
   717  
   718  	var (
   719  		enqueueIdx   int
   720  		enqueued     []byte
   721  		errTestFlush = errors.New("test flush error")
   722  	)
   723  	queue := NewMockinstanceQueue(ctrl)
   724  	queue.EXPECT().
   725  		Enqueue(gomock.Any()).
   726  		DoAndReturn(func(buf protobuf.Buffer) error {
   727  			enqueued = append(enqueued, buf.Bytes()...)
   728  			enqueueIdx++
   729  			if enqueueIdx == 1 {
   730  				return errTestFlush
   731  			}
   732  			return nil
   733  		}).
   734  		Times(2)
   735  	queue.EXPECT().Flush().MinTimes(1)
   736  	opts := testOptions()
   737  	w := newInstanceWriter(testPlacementInstance, opts).(*writer)
   738  	w.queue = queue
   739  
   740  	encoderIdx := 0
   741  	w.newLockedEncoderFn = func(protobuf.UnaggregatedOptions) *lockedEncoder {
   742  		encoder := protobuf.NewMockUnaggregatedEncoder(ctrl)
   743  		switch encoderIdx {
   744  		case 0:
   745  			encoder.EXPECT().Len().Return(0)
   746  		case 1:
   747  			encoder.EXPECT().Len().Return(2)
   748  			encoder.EXPECT().Relinquish().Return(protobuf.NewBuffer([]byte{1, 2}, nil))
   749  		case 2:
   750  			encoder.EXPECT().Len().Return(4)
   751  			encoder.EXPECT().Relinquish().Return(protobuf.NewBuffer([]byte{3, 4, 5, 6}, nil))
   752  		}
   753  		encoderIdx++
   754  		return &lockedEncoder{UnaggregatedEncoder: encoder}
   755  	}
   756  	for i := 0; i < 3; i++ {
   757  		w.encodersByShard[uint32(i)] = w.newLockedEncoderFn(opts.EncoderOptions())
   758  	}
   759  	err := w.Flush()
   760  	require.Error(t, err)
   761  	require.True(t, strings.Contains(err.Error(), errTestFlush.Error()))
   762  	sort.Slice(enqueued, func(i, j int) bool { return enqueued[i] < enqueued[j] })
   763  	require.Equal(t, []byte{1, 2, 3, 4, 5, 6}, enqueued)
   764  }
   765  
   766  func TestWriterCloseAlreadyClosed(t *testing.T) {
   767  	w := newInstanceWriter(testPlacementInstance, testOptions()).(*writer)
   768  	w.closed = true
   769  	require.Equal(t, errInstanceWriterClosed, w.Close())
   770  }
   771  
   772  func TestWriterCloseSuccess(t *testing.T) {
   773  	w := newInstanceWriter(testPlacementInstance, testOptions()).(*writer)
   774  	require.NoError(t, w.Close())
   775  }
   776  
   777  func TestWriterConcurrentWriteStress(t *testing.T) {
   778  	params := []struct {
   779  		maxInputBatchSize int
   780  		maxTimerBatchSize int
   781  		maxBatchSize      int
   782  	}{
   783  		// High likelihood of counter/gauge encoding triggering a flush in between
   784  		// releasing and re-acquiring locks when encoding large timer batches.
   785  		{
   786  			maxInputBatchSize: 150,
   787  			maxTimerBatchSize: 150,
   788  			maxBatchSize:      1000,
   789  		},
   790  		// Large timer batches.
   791  		{
   792  			maxInputBatchSize: 1000,
   793  			maxTimerBatchSize: 140,
   794  			maxBatchSize:      1440,
   795  		},
   796  	}
   797  
   798  	for _, param := range params {
   799  		testWriterConcurrentWriteStress(
   800  			t,
   801  			param.maxInputBatchSize,
   802  			param.maxTimerBatchSize,
   803  			param.maxBatchSize,
   804  		)
   805  	}
   806  }
   807  
   808  func testWriterConcurrentWriteStress(
   809  	t *testing.T,
   810  	maxInputBatchSize int,
   811  	maxTimerBatchSize int,
   812  	maxBatchSize int,
   813  ) {
   814  	ctrl := gomock.NewController(t)
   815  	defer ctrl.Finish()
   816  
   817  	var (
   818  		numIter       = 3000
   819  		shard         = uint32(0)
   820  		counters      = make([]unaggregated.Counter, numIter)
   821  		timers        = make([]unaggregated.BatchTimer, numIter)
   822  		gauges        = make([]unaggregated.Gauge, numIter)
   823  		forwarded     = make([]aggregated.ForwardedMetric, numIter)
   824  		passthroughed = make([]aggregated.Metric, numIter)
   825  		resultsLock   sync.Mutex
   826  		results       [][]byte
   827  	)
   828  
   829  	// Construct metrics input.
   830  	for i := 0; i < numIter; i++ {
   831  		counters[i] = unaggregated.Counter{
   832  			ID:    []byte(fmt.Sprintf("counter%d", i)),
   833  			Value: int64(i),
   834  		}
   835  		gauges[i] = unaggregated.Gauge{
   836  			ID:    []byte(fmt.Sprintf("gauge%d", i)),
   837  			Value: float64(i),
   838  		}
   839  		batchSize := numIter - i
   840  		if batchSize > maxInputBatchSize {
   841  			batchSize = maxInputBatchSize
   842  		}
   843  		timerVals := make([]float64, batchSize)
   844  		for j := i; j < i+batchSize; j++ {
   845  			timerVals[j-i] = float64(j)
   846  		}
   847  		timers[i] = unaggregated.BatchTimer{
   848  			ID:     []byte(fmt.Sprintf("timer%d", i)),
   849  			Values: timerVals,
   850  		}
   851  		forwardedVals := []float64{float64(i) - 0.5, float64(i), float64(i) + 0.5}
   852  		forwarded[i] = aggregated.ForwardedMetric{
   853  			Type:      metric.GaugeType,
   854  			ID:        []byte(fmt.Sprintf("forwarded%d", i)),
   855  			TimeNanos: int64(i),
   856  			Values:    forwardedVals,
   857  		}
   858  		passthroughed[i] = aggregated.Metric{
   859  			Type:      metric.GaugeType,
   860  			ID:        []byte(fmt.Sprintf("passthroughed%d", i)),
   861  			TimeNanos: int64(i),
   862  			Value:     float64(i),
   863  		}
   864  	}
   865  
   866  	queue := NewMockinstanceQueue(ctrl)
   867  	queue.EXPECT().
   868  		Enqueue(gomock.Any()).
   869  		DoAndReturn(func(buf protobuf.Buffer) error {
   870  			bytes := buf.Bytes()
   871  			cloned := make([]byte, len(bytes))
   872  			copy(cloned, bytes)
   873  			resultsLock.Lock()
   874  			results = append(results, cloned)
   875  			resultsLock.Unlock()
   876  			return nil
   877  		}).
   878  		AnyTimes()
   879  	queue.EXPECT().Flush().MinTimes(1)
   880  	opts := testOptions().
   881  		SetMaxTimerBatchSize(maxTimerBatchSize).
   882  		SetMaxBatchSize(maxBatchSize)
   883  	w := newInstanceWriter(testPlacementInstance, opts).(*writer)
   884  	w.queue = queue
   885  
   886  	var wg sync.WaitGroup
   887  	wg.Add(5)
   888  
   889  	go func() {
   890  		defer wg.Done()
   891  
   892  		for i := 0; i < numIter; i++ {
   893  			mu := unaggregated.MetricUnion{
   894  				Type:       metric.CounterType,
   895  				ID:         counters[i].ID,
   896  				CounterVal: counters[i].Value,
   897  			}
   898  			payload := payloadUnion{
   899  				payloadType: untimedType,
   900  				untimed: untimedPayload{
   901  					metric:    mu,
   902  					metadatas: testStagedMetadatas,
   903  				},
   904  			}
   905  			require.NoError(t, w.Write(shard, payload))
   906  		}
   907  	}()
   908  
   909  	go func() {
   910  		defer wg.Done()
   911  
   912  		for i := 0; i < numIter; i++ {
   913  			mu := unaggregated.MetricUnion{
   914  				Type:          metric.TimerType,
   915  				ID:            timers[i].ID,
   916  				BatchTimerVal: timers[i].Values,
   917  			}
   918  			payload := payloadUnion{
   919  				payloadType: untimedType,
   920  				untimed: untimedPayload{
   921  					metric:    mu,
   922  					metadatas: testStagedMetadatas,
   923  				},
   924  			}
   925  			require.NoError(t, w.Write(shard, payload))
   926  		}
   927  	}()
   928  
   929  	go func() {
   930  		defer wg.Done()
   931  
   932  		for i := 0; i < numIter; i++ {
   933  			mu := unaggregated.MetricUnion{
   934  				Type:     metric.GaugeType,
   935  				ID:       gauges[i].ID,
   936  				GaugeVal: gauges[i].Value,
   937  			}
   938  			payload := payloadUnion{
   939  				payloadType: untimedType,
   940  				untimed: untimedPayload{
   941  					metric:    mu,
   942  					metadatas: testStagedMetadatas,
   943  				},
   944  			}
   945  			require.NoError(t, w.Write(shard, payload))
   946  		}
   947  	}()
   948  
   949  	go func() {
   950  		defer wg.Done()
   951  
   952  		for i := 0; i < numIter; i++ {
   953  			payload := payloadUnion{
   954  				payloadType: forwardedType,
   955  				forwarded: forwardedPayload{
   956  					metric:   forwarded[i],
   957  					metadata: testForwardMetadata,
   958  				},
   959  			}
   960  			require.NoError(t, w.Write(shard, payload))
   961  		}
   962  	}()
   963  
   964  	go func() {
   965  		defer wg.Done()
   966  
   967  		for i := 0; i < numIter; i++ {
   968  			payload := payloadUnion{
   969  				payloadType: passthroughType,
   970  				passthrough: passthroughPayload{
   971  					metric:        passthroughed[i],
   972  					storagePolicy: testPassthroughMetadata,
   973  				},
   974  			}
   975  			require.NoError(t, w.Write(shard, payload))
   976  		}
   977  	}()
   978  
   979  	wg.Wait()
   980  	w.Flush()
   981  
   982  	var (
   983  		resCounters      = make([]unaggregated.Counter, 0, numIter)
   984  		resTimers        = make([]unaggregated.BatchTimer, 0, numIter)
   985  		resGauges        = make([]unaggregated.Gauge, 0, numIter)
   986  		resForwarded     = make([]aggregated.ForwardedMetric, 0, numIter)
   987  		resPassthroughed = make([]aggregated.Metric, 0, numIter)
   988  	)
   989  	for i := 0; i < len(results); i++ {
   990  		buf := bytes.NewBuffer(results[i])
   991  		iter := protobuf.NewUnaggregatedIterator(buf, protobuf.NewUnaggregatedOptions())
   992  		for iter.Next() {
   993  			msgResult := iter.Current()
   994  			switch msgResult.Type {
   995  			case encoding.CounterWithMetadatasType:
   996  				require.Equal(t, testStagedMetadatas, msgResult.CounterWithMetadatas.StagedMetadatas)
   997  				metric := cloneMetric(msgResult.CounterWithMetadatas.Counter.ToUnion())
   998  				resCounters = append(resCounters, metric.Counter())
   999  			case encoding.BatchTimerWithMetadatasType:
  1000  				require.Equal(t, testStagedMetadatas, msgResult.BatchTimerWithMetadatas.StagedMetadatas)
  1001  				metric := cloneMetric(msgResult.BatchTimerWithMetadatas.BatchTimer.ToUnion())
  1002  				resTimers = append(resTimers, metric.BatchTimer())
  1003  			case encoding.GaugeWithMetadatasType:
  1004  				require.Equal(t, testStagedMetadatas, msgResult.GaugeWithMetadatas.StagedMetadatas)
  1005  				metric := cloneMetric(msgResult.GaugeWithMetadatas.Gauge.ToUnion())
  1006  				resGauges = append(resGauges, metric.Gauge())
  1007  			case encoding.ForwardedMetricWithMetadataType:
  1008  				require.Equal(t, testForwardMetadata, msgResult.ForwardedMetricWithMetadata.ForwardMetadata)
  1009  				metric := cloneForwardedMetric(msgResult.ForwardedMetricWithMetadata.ForwardedMetric)
  1010  				resForwarded = append(resForwarded, metric)
  1011  			case encoding.PassthroughMetricWithMetadataType:
  1012  				require.Equal(t, testPassthroughMetadata, msgResult.PassthroughMetricWithMetadata.StoragePolicy)
  1013  				metric := clonePassthroughedMetric(msgResult.PassthroughMetricWithMetadata.Metric)
  1014  				resPassthroughed = append(resPassthroughed, metric)
  1015  			default:
  1016  				require.Fail(t, "unrecognized message type %v", msgResult.Type)
  1017  			}
  1018  		}
  1019  		require.Equal(t, io.EOF, iter.Err())
  1020  	}
  1021  
  1022  	// Sort counters for comparison purposes.
  1023  	sort.Slice(counters, func(i, j int) bool {
  1024  		return bytes.Compare(counters[i].ID, counters[j].ID) < 0
  1025  	})
  1026  	sort.Slice(resCounters, func(i, j int) bool {
  1027  		return bytes.Compare(resCounters[i].ID, resCounters[j].ID) < 0
  1028  	})
  1029  	require.Equal(t, counters, resCounters)
  1030  
  1031  	// Sort timers for comparison purposes.
  1032  	sort.Slice(timers, func(i, j int) bool {
  1033  		return bytes.Compare(timers[i].ID, timers[j].ID) < 0
  1034  	})
  1035  	sort.Slice(resTimers, func(i, j int) bool {
  1036  		return bytes.Compare(resTimers[i].ID, resTimers[j].ID) < 0
  1037  	})
  1038  	// Merge timers if necessary for comparison since they may be split into multiple batches.
  1039  	mergedResTimers := make([]unaggregated.BatchTimer, 0, numIter)
  1040  	curr := 0
  1041  	for i := 0; i < len(resTimers); i++ {
  1042  		if bytes.Equal(resTimers[curr].ID, resTimers[i].ID) {
  1043  			continue
  1044  		}
  1045  		var mergedValues []float64
  1046  		for j := curr; j < i; j++ {
  1047  			mergedValues = append(mergedValues, resTimers[j].Values...)
  1048  		}
  1049  		sort.Float64s(mergedValues)
  1050  		mergedResTimers = append(mergedResTimers, unaggregated.BatchTimer{
  1051  			ID:     resTimers[curr].ID,
  1052  			Values: mergedValues,
  1053  		})
  1054  		curr = i
  1055  	}
  1056  	if curr < len(resTimers) {
  1057  		var mergedValues []float64
  1058  		for j := curr; j < len(resTimers); j++ {
  1059  			mergedValues = append(mergedValues, resTimers[j].Values...)
  1060  		}
  1061  		sort.Float64s(mergedValues)
  1062  		mergedResTimers = append(mergedResTimers, unaggregated.BatchTimer{
  1063  			ID:     resTimers[curr].ID,
  1064  			Values: mergedValues,
  1065  		})
  1066  	}
  1067  	require.Equal(t, timers, mergedResTimers)
  1068  
  1069  	// Sort gauges for comparison purposes.
  1070  	sort.Slice(gauges, func(i, j int) bool {
  1071  		return bytes.Compare(gauges[i].ID, gauges[j].ID) < 0
  1072  	})
  1073  	sort.Slice(resGauges, func(i, j int) bool {
  1074  		return bytes.Compare(resGauges[i].ID, resGauges[j].ID) < 0
  1075  	})
  1076  	require.Equal(t, gauges, resGauges)
  1077  
  1078  	// Sort forwarded for comparison purposes.
  1079  	sort.Slice(forwarded, func(i, j int) bool {
  1080  		return bytes.Compare(forwarded[i].ID, forwarded[j].ID) < 0
  1081  	})
  1082  	sort.Slice(resForwarded, func(i, j int) bool {
  1083  		return bytes.Compare(resForwarded[i].ID, resForwarded[j].ID) < 0
  1084  	})
  1085  	require.Equal(t, forwarded, resForwarded)
  1086  }
  1087  
  1088  func TestRefCountedWriter(t *testing.T) {
  1089  	opts := testOptions()
  1090  	w := newRefCountedWriter(testPlacementInstance, opts)
  1091  	w.IncRef()
  1092  
  1093  	require.False(t, w.instanceWriter.(*writer).closed)
  1094  	w.DecRef()
  1095  	require.True(t, clock.WaitUntil(func() bool {
  1096  		wr := w.instanceWriter.(*writer)
  1097  		wr.Lock()
  1098  		defer wr.Unlock()
  1099  		return wr.closed
  1100  	}, 3*time.Second))
  1101  }
  1102  
  1103  func cloneMetric(m unaggregated.MetricUnion) unaggregated.MetricUnion {
  1104  	mu := m
  1105  	clonedID := make(id.RawID, len(m.ID))
  1106  	copy(clonedID, m.ID)
  1107  	mu.ID = clonedID
  1108  	if m.Type == metric.TimerType {
  1109  		clonedTimerVal := make([]float64, len(m.BatchTimerVal))
  1110  		copy(clonedTimerVal, m.BatchTimerVal)
  1111  		mu.BatchTimerVal = clonedTimerVal
  1112  	}
  1113  	return mu
  1114  }
  1115  
  1116  func cloneForwardedMetric(m aggregated.ForwardedMetric) aggregated.ForwardedMetric {
  1117  	cloned := m
  1118  	cloned.ID = append([]byte(nil), m.ID...)
  1119  	cloned.Values = append([]float64(nil), m.Values...)
  1120  	return cloned
  1121  }
  1122  
  1123  func clonePassthroughedMetric(m aggregated.Metric) aggregated.Metric {
  1124  	cloned := m
  1125  	cloned.ID = append([]byte(nil), m.ID...)
  1126  	cloned.Value = m.Value
  1127  	return cloned
  1128  }