github.com/Hnampk/fabric@v2.1.1+incompatible/gossip/state/state_test.go (about)

     1  /*
     2  Copyright IBM Corp. All Rights Reserved.
     3  
     4  SPDX-License-Identifier: Apache-2.0
     5  */
     6  
     7  package state
     8  
     9  import (
    10  	"bytes"
    11  	"errors"
    12  	"fmt"
    13  	"math/rand"
    14  	"net"
    15  	"sync"
    16  	"sync/atomic"
    17  	"testing"
    18  	"time"
    19  
    20  	pb "github.com/golang/protobuf/proto"
    21  	pcomm "github.com/hyperledger/fabric-protos-go/common"
    22  	proto "github.com/hyperledger/fabric-protos-go/gossip"
    23  	"github.com/hyperledger/fabric-protos-go/ledger/rwset"
    24  	tspb "github.com/hyperledger/fabric-protos-go/transientstore"
    25  	"github.com/hyperledger/fabric/bccsp/factory"
    26  	"github.com/hyperledger/fabric/common/configtx/test"
    27  	errors2 "github.com/hyperledger/fabric/common/errors"
    28  	"github.com/hyperledger/fabric/common/flogging"
    29  	"github.com/hyperledger/fabric/common/metrics/disabled"
    30  	"github.com/hyperledger/fabric/core/committer"
    31  	"github.com/hyperledger/fabric/core/committer/txvalidator"
    32  	"github.com/hyperledger/fabric/core/ledger"
    33  	"github.com/hyperledger/fabric/core/mocks/validator"
    34  	"github.com/hyperledger/fabric/core/transientstore"
    35  	"github.com/hyperledger/fabric/gossip/api"
    36  	"github.com/hyperledger/fabric/gossip/comm"
    37  	"github.com/hyperledger/fabric/gossip/common"
    38  	"github.com/hyperledger/fabric/gossip/discovery"
    39  	"github.com/hyperledger/fabric/gossip/gossip"
    40  	"github.com/hyperledger/fabric/gossip/gossip/algo"
    41  	"github.com/hyperledger/fabric/gossip/gossip/channel"
    42  	"github.com/hyperledger/fabric/gossip/metrics"
    43  	"github.com/hyperledger/fabric/gossip/privdata"
    44  	capabilitymock "github.com/hyperledger/fabric/gossip/privdata/mocks"
    45  	"github.com/hyperledger/fabric/gossip/protoext"
    46  	"github.com/hyperledger/fabric/gossip/state/mocks"
    47  	gossiputil "github.com/hyperledger/fabric/gossip/util"
    48  	gutil "github.com/hyperledger/fabric/gossip/util"
    49  	corecomm "github.com/hyperledger/fabric/internal/pkg/comm"
    50  	"github.com/hyperledger/fabric/protoutil"
    51  	"github.com/onsi/gomega"
    52  	"github.com/onsi/gomega/gbytes"
    53  	"github.com/stretchr/testify/assert"
    54  	"github.com/stretchr/testify/mock"
    55  	"go.uber.org/zap"
    56  	"go.uber.org/zap/zapcore"
    57  )
    58  
    59  var (
    60  	orgID = []byte("ORG1")
    61  
    62  	noopPeerIdentityAcceptor = func(identity api.PeerIdentityType) error {
    63  		return nil
    64  	}
    65  )
    66  
    67  type peerIdentityAcceptor func(identity api.PeerIdentityType) error
    68  
    69  type joinChanMsg struct {
    70  }
    71  
    72  func init() {
    73  	gutil.SetupTestLogging()
    74  	factory.InitFactories(nil)
    75  }
    76  
    77  // SequenceNumber returns the sequence number of the block that the message
    78  // is derived from
    79  func (*joinChanMsg) SequenceNumber() uint64 {
    80  	return uint64(time.Now().UnixNano())
    81  }
    82  
    83  // Members returns the organizations of the channel
    84  func (jcm *joinChanMsg) Members() []api.OrgIdentityType {
    85  	return []api.OrgIdentityType{orgID}
    86  }
    87  
    88  // AnchorPeersOf returns the anchor peers of the given organization
    89  func (jcm *joinChanMsg) AnchorPeersOf(org api.OrgIdentityType) []api.AnchorPeer {
    90  	return []api.AnchorPeer{}
    91  }
    92  
    93  type orgCryptoService struct {
    94  }
    95  
    96  // OrgByPeerIdentity returns the OrgIdentityType
    97  // of a given peer identity
    98  func (*orgCryptoService) OrgByPeerIdentity(identity api.PeerIdentityType) api.OrgIdentityType {
    99  	return orgID
   100  }
   101  
   102  // Verify verifies a JoinChannelMessage, returns nil on success,
   103  // and an error on failure
   104  func (*orgCryptoService) Verify(joinChanMsg api.JoinChannelMessage) error {
   105  	return nil
   106  }
   107  
   108  type cryptoServiceMock struct {
   109  	acceptor peerIdentityAcceptor
   110  }
   111  
   112  func (cryptoServiceMock) Expiration(peerIdentity api.PeerIdentityType) (time.Time, error) {
   113  	return time.Now().Add(time.Hour), nil
   114  }
   115  
   116  // GetPKIidOfCert returns the PKI-ID of a peer's identity
   117  func (*cryptoServiceMock) GetPKIidOfCert(peerIdentity api.PeerIdentityType) common.PKIidType {
   118  	return common.PKIidType(peerIdentity)
   119  }
   120  
   121  // VerifyBlock returns nil if the block is properly signed,
   122  // else returns error
   123  func (*cryptoServiceMock) VerifyBlock(channelID common.ChannelID, seqNum uint64, signedBlock *pcomm.Block) error {
   124  	return nil
   125  }
   126  
   127  // Sign signs msg with this peer's signing key and outputs
   128  // the signature if no error occurred.
   129  func (*cryptoServiceMock) Sign(msg []byte) ([]byte, error) {
   130  	clone := make([]byte, len(msg))
   131  	copy(clone, msg)
   132  	return clone, nil
   133  }
   134  
   135  // Verify checks that signature is a valid signature of message under a peer's verification key.
   136  // If the verification succeeded, Verify returns nil meaning no error occurred.
   137  // If peerCert is nil, then the signature is verified against this peer's verification key.
   138  func (*cryptoServiceMock) Verify(peerIdentity api.PeerIdentityType, signature, message []byte) error {
   139  	equal := bytes.Equal(signature, message)
   140  	if !equal {
   141  		return fmt.Errorf("Wrong signature:%v, %v", signature, message)
   142  	}
   143  	return nil
   144  }
   145  
   146  // VerifyByChannel checks that signature is a valid signature of message
   147  // under a peer's verification key, but also in the context of a specific channel.
   148  // If the verification succeeded, Verify returns nil meaning no error occurred.
   149  // If peerIdentity is nil, then the signature is verified against this peer's verification key.
   150  func (cs *cryptoServiceMock) VerifyByChannel(channelID common.ChannelID, peerIdentity api.PeerIdentityType, signature, message []byte) error {
   151  	return cs.acceptor(peerIdentity)
   152  }
   153  
   154  func (*cryptoServiceMock) ValidateIdentity(peerIdentity api.PeerIdentityType) error {
   155  	return nil
   156  }
   157  
   158  func bootPeersWithPorts(ports ...int) []string {
   159  	var peers []string
   160  	for _, port := range ports {
   161  		peers = append(peers, fmt.Sprintf("127.0.0.1:%d", port))
   162  	}
   163  	return peers
   164  }
   165  
   166  type peerNodeGossipSupport interface {
   167  	GossipAdapter
   168  	Stop()
   169  	JoinChan(joinMsg api.JoinChannelMessage, channelID common.ChannelID)
   170  }
   171  
   172  // Simple presentation of peer which includes only
   173  // communication module, gossip and state transfer
   174  type peerNode struct {
   175  	port   int
   176  	g      peerNodeGossipSupport
   177  	s      *GossipStateProviderImpl
   178  	cs     *cryptoServiceMock
   179  	commit committer.Committer
   180  	grpc   *corecomm.GRPCServer
   181  }
   182  
   183  // Shutting down all modules used
   184  func (node *peerNode) shutdown() {
   185  	node.s.Stop()
   186  	node.g.Stop()
   187  	node.grpc.Stop()
   188  }
   189  
   190  type mockCommitter struct {
   191  	*mock.Mock
   192  	sync.Mutex
   193  }
   194  
   195  func (mc *mockCommitter) GetConfigHistoryRetriever() (ledger.ConfigHistoryRetriever, error) {
   196  	args := mc.Called()
   197  	return args.Get(0).(ledger.ConfigHistoryRetriever), args.Error(1)
   198  }
   199  
   200  func (mc *mockCommitter) GetPvtDataByNum(blockNum uint64, filter ledger.PvtNsCollFilter) ([]*ledger.TxPvtData, error) {
   201  	args := mc.Called(blockNum, filter)
   202  	return args.Get(0).([]*ledger.TxPvtData), args.Error(1)
   203  }
   204  
   205  func (mc *mockCommitter) CommitLegacy(blockAndPvtData *ledger.BlockAndPvtData, commitOpts *ledger.CommitOptions) error {
   206  	mc.Lock()
   207  	m := mc.Mock
   208  	mc.Unlock()
   209  	m.Called(blockAndPvtData.Block)
   210  	return nil
   211  }
   212  
   213  func (mc *mockCommitter) GetPvtDataAndBlockByNum(seqNum uint64) (*ledger.BlockAndPvtData, error) {
   214  	mc.Lock()
   215  	m := mc.Mock
   216  	mc.Unlock()
   217  
   218  	args := m.Called(seqNum)
   219  	return args.Get(0).(*ledger.BlockAndPvtData), args.Error(1)
   220  }
   221  
   222  func (mc *mockCommitter) LedgerHeight() (uint64, error) {
   223  	mc.Lock()
   224  	m := mc.Mock
   225  	mc.Unlock()
   226  	args := m.Called()
   227  	if args.Get(1) == nil {
   228  		return args.Get(0).(uint64), nil
   229  	}
   230  	return args.Get(0).(uint64), args.Get(1).(error)
   231  }
   232  
   233  func (mc *mockCommitter) DoesPvtDataInfoExistInLedger(blkNum uint64) (bool, error) {
   234  	mc.Lock()
   235  	m := mc.Mock
   236  	mc.Unlock()
   237  	args := m.Called(blkNum)
   238  	return args.Get(0).(bool), args.Error(1)
   239  }
   240  
   241  func (mc *mockCommitter) GetBlocks(blockSeqs []uint64) []*pcomm.Block {
   242  	mc.Lock()
   243  	m := mc.Mock
   244  	mc.Unlock()
   245  
   246  	if m.Called(blockSeqs).Get(0) == nil {
   247  		return nil
   248  	}
   249  	return m.Called(blockSeqs).Get(0).([]*pcomm.Block)
   250  }
   251  
   252  func (*mockCommitter) GetMissingPvtDataTracker() (ledger.MissingPvtDataTracker, error) {
   253  	panic("implement me")
   254  }
   255  
   256  func (*mockCommitter) CommitPvtDataOfOldBlocks(reconciledPvtdata []*ledger.ReconciledPvtdata) ([]*ledger.PvtdataHashMismatch, error) {
   257  	panic("implement me")
   258  }
   259  
   260  func (*mockCommitter) Close() {
   261  }
   262  
   263  type ramLedger struct {
   264  	ledger map[uint64]*ledger.BlockAndPvtData
   265  	sync.RWMutex
   266  }
   267  
   268  func (mock *ramLedger) GetMissingPvtDataTracker() (ledger.MissingPvtDataTracker, error) {
   269  	panic("implement me")
   270  }
   271  
   272  func (mock *ramLedger) CommitPvtDataOfOldBlocks(reconciledPvtdata []*ledger.ReconciledPvtdata) ([]*ledger.PvtdataHashMismatch, error) {
   273  	panic("implement me")
   274  }
   275  
   276  func (mock *ramLedger) GetConfigHistoryRetriever() (ledger.ConfigHistoryRetriever, error) {
   277  	panic("implement me")
   278  }
   279  
   280  func (mock *ramLedger) GetPvtDataAndBlockByNum(blockNum uint64, filter ledger.PvtNsCollFilter) (*ledger.BlockAndPvtData, error) {
   281  	mock.RLock()
   282  	defer mock.RUnlock()
   283  
   284  	if block, ok := mock.ledger[blockNum]; !ok {
   285  		return nil, errors.New(fmt.Sprintf("no block with seq = %d found", blockNum))
   286  	} else {
   287  		return block, nil
   288  	}
   289  }
   290  
   291  func (mock *ramLedger) GetPvtDataByNum(blockNum uint64, filter ledger.PvtNsCollFilter) ([]*ledger.TxPvtData, error) {
   292  	panic("implement me")
   293  }
   294  
   295  func (mock *ramLedger) CommitLegacy(blockAndPvtdata *ledger.BlockAndPvtData, commitOpts *ledger.CommitOptions) error {
   296  	mock.Lock()
   297  	defer mock.Unlock()
   298  
   299  	if blockAndPvtdata != nil && blockAndPvtdata.Block != nil {
   300  		mock.ledger[blockAndPvtdata.Block.Header.Number] = blockAndPvtdata
   301  		return nil
   302  	}
   303  	return errors.New("invalid input parameters for block and private data param")
   304  }
   305  
   306  func (mock *ramLedger) GetBlockchainInfo() (*pcomm.BlockchainInfo, error) {
   307  	mock.RLock()
   308  	defer mock.RUnlock()
   309  
   310  	currentBlock := mock.ledger[uint64(len(mock.ledger)-1)].Block
   311  	return &pcomm.BlockchainInfo{
   312  		Height:            currentBlock.Header.Number + 1,
   313  		CurrentBlockHash:  protoutil.BlockHeaderHash(currentBlock.Header),
   314  		PreviousBlockHash: currentBlock.Header.PreviousHash,
   315  	}, nil
   316  }
   317  
   318  func (mock *ramLedger) DoesPvtDataInfoExist(blkNum uint64) (bool, error) {
   319  	return false, nil
   320  }
   321  
   322  func (mock *ramLedger) GetBlockByNumber(blockNumber uint64) (*pcomm.Block, error) {
   323  	mock.RLock()
   324  	defer mock.RUnlock()
   325  
   326  	if blockAndPvtData, ok := mock.ledger[blockNumber]; !ok {
   327  		return nil, errors.New(fmt.Sprintf("no block with seq = %d found", blockNumber))
   328  	} else {
   329  		return blockAndPvtData.Block, nil
   330  	}
   331  }
   332  
   333  func (mock *ramLedger) Close() {
   334  
   335  }
   336  
   337  // Create new instance of KVLedger to be used for testing
   338  func newCommitter() committer.Committer {
   339  	cb, _ := test.MakeGenesisBlock("testChain")
   340  	ldgr := &ramLedger{
   341  		ledger: make(map[uint64]*ledger.BlockAndPvtData),
   342  	}
   343  	ldgr.CommitLegacy(&ledger.BlockAndPvtData{Block: cb}, &ledger.CommitOptions{})
   344  	return committer.NewLedgerCommitter(ldgr)
   345  }
   346  
   347  func newPeerNodeWithGossip(id int, committer committer.Committer,
   348  	acceptor peerIdentityAcceptor, g peerNodeGossipSupport, bootPorts ...int) *peerNode {
   349  	logger := flogging.MustGetLogger(gutil.StateLogger)
   350  	return newPeerNodeWithGossipWithValidator(logger, id, committer, acceptor, g, &validator.MockValidator{}, bootPorts...)
   351  }
   352  
   353  // Constructing pseudo peer node, simulating only gossip and state transfer part
   354  func newPeerNodeWithGossipWithValidatorWithMetrics(logger gutil.Logger, id int, committer committer.Committer,
   355  	acceptor peerIdentityAcceptor, g peerNodeGossipSupport, v txvalidator.Validator,
   356  	gossipMetrics *metrics.GossipMetrics, bootPorts ...int) (node *peerNode, port int) {
   357  	cs := &cryptoServiceMock{acceptor: acceptor}
   358  	port, gRPCServer, certs, secureDialOpts, _ := gossiputil.CreateGRPCLayer()
   359  
   360  	if g == nil {
   361  		config := &gossip.Config{
   362  			BindPort:                     port,
   363  			BootstrapPeers:               bootPeersWithPorts(bootPorts...),
   364  			ID:                           fmt.Sprintf("p%d", id),
   365  			MaxBlockCountToStore:         0,
   366  			MaxPropagationBurstLatency:   time.Duration(10) * time.Millisecond,
   367  			MaxPropagationBurstSize:      10,
   368  			PropagateIterations:          1,
   369  			PropagatePeerNum:             3,
   370  			PullInterval:                 time.Duration(4) * time.Second,
   371  			PullPeerNum:                  5,
   372  			InternalEndpoint:             fmt.Sprintf("127.0.0.1:%d", port),
   373  			PublishCertPeriod:            10 * time.Second,
   374  			RequestStateInfoInterval:     4 * time.Second,
   375  			PublishStateInfoInterval:     4 * time.Second,
   376  			TimeForMembershipTracker:     5 * time.Second,
   377  			TLSCerts:                     certs,
   378  			DigestWaitTime:               algo.DefDigestWaitTime,
   379  			RequestWaitTime:              algo.DefRequestWaitTime,
   380  			ResponseWaitTime:             algo.DefResponseWaitTime,
   381  			DialTimeout:                  comm.DefDialTimeout,
   382  			ConnTimeout:                  comm.DefConnTimeout,
   383  			RecvBuffSize:                 comm.DefRecvBuffSize,
   384  			SendBuffSize:                 comm.DefSendBuffSize,
   385  			MsgExpirationTimeout:         channel.DefMsgExpirationTimeout,
   386  			AliveTimeInterval:            discovery.DefAliveTimeInterval,
   387  			AliveExpirationTimeout:       discovery.DefAliveExpirationTimeout,
   388  			AliveExpirationCheckInterval: discovery.DefAliveExpirationCheckInterval,
   389  			ReconnectInterval:            discovery.DefReconnectInterval,
   390  		}
   391  
   392  		selfID := api.PeerIdentityType(config.InternalEndpoint)
   393  		mcs := &cryptoServiceMock{acceptor: noopPeerIdentityAcceptor}
   394  		g = gossip.New(config, gRPCServer.Server(), &orgCryptoService{}, mcs, selfID, secureDialOpts, gossipMetrics)
   395  	}
   396  
   397  	g.JoinChan(&joinChanMsg{}, common.ChannelID("testchannelid"))
   398  
   399  	go func() {
   400  		gRPCServer.Start()
   401  	}()
   402  
   403  	// Initialize pseudo peer simulator, which has only three
   404  	// basic parts
   405  
   406  	servicesAdapater := &ServicesMediator{GossipAdapter: g, MCSAdapter: cs}
   407  	coordConfig := privdata.CoordinatorConfig{
   408  		PullRetryThreshold:             0,
   409  		TransientBlockRetention:        1000,
   410  		SkipPullingInvalidTransactions: false,
   411  	}
   412  
   413  	mspID := "Org1MSP"
   414  	capabilityProvider := &capabilitymock.CapabilityProvider{}
   415  	appCapability := &capabilitymock.AppCapabilities{}
   416  	capabilityProvider.On("Capabilities").Return(appCapability)
   417  	appCapability.On("StorePvtDataOfInvalidTx").Return(true)
   418  	coord := privdata.NewCoordinator(mspID, privdata.Support{
   419  		Validator:          v,
   420  		Committer:          committer,
   421  		CapabilityProvider: capabilityProvider,
   422  	}, &transientstore.Store{}, protoutil.SignedData{}, gossipMetrics.PrivdataMetrics, coordConfig, nil)
   423  	stateConfig := &StateConfig{
   424  		StateCheckInterval:   DefStateCheckInterval,
   425  		StateResponseTimeout: DefStateResponseTimeout,
   426  		StateBatchSize:       DefStateBatchSize,
   427  		StateMaxRetries:      DefStateMaxRetries,
   428  		StateBlockBufferSize: DefStateBlockBufferSize,
   429  		StateChannelSize:     DefStateChannelSize,
   430  		StateEnabled:         DefStateEnabled,
   431  	}
   432  	sp := NewGossipStateProvider(logger, "testchannelid", servicesAdapater, coord, gossipMetrics.StateMetrics, blocking, stateConfig)
   433  	if sp == nil {
   434  		gRPCServer.Stop()
   435  		return nil, port
   436  	}
   437  
   438  	return &peerNode{
   439  		port:   port,
   440  		g:      g,
   441  		s:      sp.(*GossipStateProviderImpl),
   442  		commit: committer,
   443  		cs:     cs,
   444  		grpc:   gRPCServer,
   445  	}, port
   446  
   447  }
   448  
   449  // add metrics provider for metrics testing
   450  func newPeerNodeWithGossipWithMetrics(id int, committer committer.Committer,
   451  	acceptor peerIdentityAcceptor, g peerNodeGossipSupport, gossipMetrics *metrics.GossipMetrics) *peerNode {
   452  	logger := flogging.MustGetLogger(gutil.StateLogger)
   453  	node, _ := newPeerNodeWithGossipWithValidatorWithMetrics(logger, id, committer, acceptor, g,
   454  		&validator.MockValidator{}, gossipMetrics)
   455  	return node
   456  }
   457  
   458  // Constructing pseudo peer node, simulating only gossip and state transfer part
   459  func newPeerNodeWithGossipWithValidator(logger gutil.Logger, id int, committer committer.Committer,
   460  	acceptor peerIdentityAcceptor, g peerNodeGossipSupport, v txvalidator.Validator, bootPorts ...int) *peerNode {
   461  	gossipMetrics := metrics.NewGossipMetrics(&disabled.Provider{})
   462  	node, _ := newPeerNodeWithGossipWithValidatorWithMetrics(logger, id, committer, acceptor, g, v, gossipMetrics, bootPorts...)
   463  	return node
   464  }
   465  
   466  // Constructing pseudo peer node, simulating only gossip and state transfer part
   467  func newPeerNode(id int, committer committer.Committer, acceptor peerIdentityAcceptor, bootPorts ...int) *peerNode {
   468  	return newPeerNodeWithGossip(id, committer, acceptor, nil, bootPorts...)
   469  }
   470  
   471  // Constructing pseudo boot node, simulating only gossip and state transfer part, return port
   472  func newBootNode(id int, committer committer.Committer, acceptor peerIdentityAcceptor) (node *peerNode, port int) {
   473  	v := &validator.MockValidator{}
   474  	gossipMetrics := metrics.NewGossipMetrics(&disabled.Provider{})
   475  	logger := flogging.MustGetLogger(gutil.StateLogger)
   476  	return newPeerNodeWithGossipWithValidatorWithMetrics(logger, id, committer, acceptor, nil, v, gossipMetrics)
   477  }
   478  
   479  func TestNilDirectMsg(t *testing.T) {
   480  	t.Parallel()
   481  	mc := &mockCommitter{Mock: &mock.Mock{}}
   482  	mc.On("LedgerHeight", mock.Anything).Return(uint64(1), nil)
   483  	g := &mocks.GossipMock{}
   484  	g.On("Accept", mock.Anything, false).Return(make(<-chan *proto.GossipMessage), nil)
   485  	g.On("Accept", mock.Anything, true).Return(nil, make(chan protoext.ReceivedMessage))
   486  	p := newPeerNodeWithGossip(0, mc, noopPeerIdentityAcceptor, g)
   487  	defer p.shutdown()
   488  	p.s.handleStateRequest(nil)
   489  	p.s.directMessage(nil)
   490  	sMsg, _ := protoext.NoopSign(p.s.stateRequestMessage(uint64(10), uint64(8)))
   491  	req := &comm.ReceivedMessageImpl{
   492  		SignedGossipMessage: sMsg,
   493  	}
   494  	p.s.directMessage(req)
   495  }
   496  
   497  func TestNilAddPayload(t *testing.T) {
   498  	t.Parallel()
   499  	mc := &mockCommitter{Mock: &mock.Mock{}}
   500  	mc.On("LedgerHeight", mock.Anything).Return(uint64(1), nil)
   501  	g := &mocks.GossipMock{}
   502  	g.On("Accept", mock.Anything, false).Return(make(<-chan *proto.GossipMessage), nil)
   503  	g.On("Accept", mock.Anything, true).Return(nil, make(chan protoext.ReceivedMessage))
   504  	p := newPeerNodeWithGossip(0, mc, noopPeerIdentityAcceptor, g)
   505  	defer p.shutdown()
   506  	err := p.s.AddPayload(nil)
   507  	assert.Error(t, err)
   508  	assert.Contains(t, err.Error(), "nil")
   509  }
   510  
   511  func TestAddPayloadLedgerUnavailable(t *testing.T) {
   512  	t.Parallel()
   513  	mc := &mockCommitter{Mock: &mock.Mock{}}
   514  	mc.On("LedgerHeight", mock.Anything).Return(uint64(1), nil)
   515  	g := &mocks.GossipMock{}
   516  	g.On("Accept", mock.Anything, false).Return(make(<-chan *proto.GossipMessage), nil)
   517  	g.On("Accept", mock.Anything, true).Return(nil, make(chan protoext.ReceivedMessage))
   518  	p := newPeerNodeWithGossip(0, mc, noopPeerIdentityAcceptor, g)
   519  	defer p.shutdown()
   520  	// Simulate a problem in the ledger
   521  	failedLedger := mock.Mock{}
   522  	failedLedger.On("LedgerHeight", mock.Anything).Return(uint64(0), errors.New("cannot query ledger"))
   523  	mc.Lock()
   524  	mc.Mock = &failedLedger
   525  	mc.Unlock()
   526  
   527  	rawblock := protoutil.NewBlock(uint64(1), []byte{})
   528  	b, _ := pb.Marshal(rawblock)
   529  	err := p.s.AddPayload(&proto.Payload{
   530  		SeqNum: uint64(1),
   531  		Data:   b,
   532  	})
   533  	assert.Error(t, err)
   534  	assert.Contains(t, err.Error(), "Failed obtaining ledger height")
   535  	assert.Contains(t, err.Error(), "cannot query ledger")
   536  }
   537  
   538  func TestLargeBlockGap(t *testing.T) {
   539  	// Scenario: the peer knows of a peer who has a ledger height much higher
   540  	// than itself (500 blocks higher).
   541  	// The peer needs to ask blocks in a way such that the size of the payload buffer
   542  	// never rises above a certain threshold.
   543  	t.Parallel()
   544  	mc := &mockCommitter{Mock: &mock.Mock{}}
   545  	blocksPassedToLedger := make(chan uint64, 200)
   546  	mc.On("CommitLegacy", mock.Anything).Run(func(arg mock.Arguments) {
   547  		blocksPassedToLedger <- arg.Get(0).(*pcomm.Block).Header.Number
   548  	})
   549  	msgsFromPeer := make(chan protoext.ReceivedMessage)
   550  	mc.On("LedgerHeight", mock.Anything).Return(uint64(1), nil)
   551  	mc.On("DoesPvtDataInfoExistInLedger", mock.Anything).Return(false, nil)
   552  	g := &mocks.GossipMock{}
   553  	membership := []discovery.NetworkMember{
   554  		{
   555  			PKIid:    common.PKIidType("a"),
   556  			Endpoint: "a",
   557  			Properties: &proto.Properties{
   558  				LedgerHeight: 500,
   559  			},
   560  		}}
   561  	g.On("PeersOfChannel", mock.Anything).Return(membership)
   562  	g.On("Accept", mock.Anything, false).Return(make(<-chan *proto.GossipMessage), nil)
   563  	g.On("Accept", mock.Anything, true).Return(nil, msgsFromPeer)
   564  	g.On("Send", mock.Anything, mock.Anything).Run(func(arguments mock.Arguments) {
   565  		msg := arguments.Get(0).(*proto.GossipMessage)
   566  		// The peer requested a state request
   567  		req := msg.GetStateRequest()
   568  		// Construct a skeleton for the response
   569  		res := &proto.GossipMessage{
   570  			Nonce:   msg.Nonce,
   571  			Channel: []byte("testchannelid"),
   572  			Content: &proto.GossipMessage_StateResponse{
   573  				StateResponse: &proto.RemoteStateResponse{},
   574  			},
   575  		}
   576  		// Populate the response with payloads according to what the peer asked
   577  		for seq := req.StartSeqNum; seq <= req.EndSeqNum; seq++ {
   578  			rawblock := protoutil.NewBlock(seq, []byte{})
   579  			b, _ := pb.Marshal(rawblock)
   580  			payload := &proto.Payload{
   581  				SeqNum: seq,
   582  				Data:   b,
   583  			}
   584  			res.GetStateResponse().Payloads = append(res.GetStateResponse().Payloads, payload)
   585  		}
   586  		// Finally, send the response down the channel the peer expects to receive it from
   587  		sMsg, _ := protoext.NoopSign(res)
   588  		msgsFromPeer <- &comm.ReceivedMessageImpl{
   589  			SignedGossipMessage: sMsg,
   590  		}
   591  	})
   592  	p := newPeerNodeWithGossip(0, mc, noopPeerIdentityAcceptor, g)
   593  	defer p.shutdown()
   594  
   595  	// Process blocks at a speed of 20 Millisecond for each block.
   596  	// The imaginative peer that responds to state
   597  	// If the payload buffer expands above defMaxBlockDistance*2 + defAntiEntropyBatchSize blocks, fail the test
   598  	blockProcessingTime := 20 * time.Millisecond // 10 seconds for total 500 blocks
   599  	expectedSequence := 1
   600  	for expectedSequence < 500 {
   601  		blockSeq := <-blocksPassedToLedger
   602  		assert.Equal(t, expectedSequence, int(blockSeq))
   603  		// Ensure payload buffer isn't over-populated
   604  		assert.True(t, p.s.payloads.Size() <= defMaxBlockDistance*2+defAntiEntropyBatchSize, "payload buffer size is %d", p.s.payloads.Size())
   605  		expectedSequence++
   606  		time.Sleep(blockProcessingTime)
   607  	}
   608  }
   609  
   610  func TestOverPopulation(t *testing.T) {
   611  	// Scenario: Add to the state provider blocks
   612  	// with a gap in between, and ensure that the payload buffer
   613  	// rejects blocks starting if the distance between the ledger height to the latest
   614  	// block it contains is bigger than defMaxBlockDistance.
   615  	t.Parallel()
   616  	mc := &mockCommitter{Mock: &mock.Mock{}}
   617  	blocksPassedToLedger := make(chan uint64, 10)
   618  	mc.On("CommitLegacy", mock.Anything).Run(func(arg mock.Arguments) {
   619  		blocksPassedToLedger <- arg.Get(0).(*pcomm.Block).Header.Number
   620  	})
   621  	mc.On("LedgerHeight", mock.Anything).Return(uint64(1), nil)
   622  	mc.On("DoesPvtDataInfoExistInLedger", mock.Anything).Return(false, nil)
   623  	g := &mocks.GossipMock{}
   624  	g.On("Accept", mock.Anything, false).Return(make(<-chan *proto.GossipMessage), nil)
   625  	g.On("Accept", mock.Anything, true).Return(nil, make(chan protoext.ReceivedMessage))
   626  	p := newPeerNode(0, mc, noopPeerIdentityAcceptor)
   627  	defer p.shutdown()
   628  
   629  	// Add some blocks in a sequential manner and make sure it works
   630  	for i := 1; i <= 4; i++ {
   631  		rawblock := protoutil.NewBlock(uint64(i), []byte{})
   632  		b, _ := pb.Marshal(rawblock)
   633  		assert.NoError(t, p.s.addPayload(&proto.Payload{
   634  			SeqNum: uint64(i),
   635  			Data:   b,
   636  		}, nonBlocking))
   637  	}
   638  
   639  	// Add payloads from 10 to defMaxBlockDistance, while we're missing blocks [5,9]
   640  	// Should succeed
   641  	for i := 10; i <= defMaxBlockDistance; i++ {
   642  		rawblock := protoutil.NewBlock(uint64(i), []byte{})
   643  		b, _ := pb.Marshal(rawblock)
   644  		assert.NoError(t, p.s.addPayload(&proto.Payload{
   645  			SeqNum: uint64(i),
   646  			Data:   b,
   647  		}, nonBlocking))
   648  	}
   649  
   650  	// Add payloads from defMaxBlockDistance + 2 to defMaxBlockDistance * 10
   651  	// Should fail.
   652  	for i := defMaxBlockDistance + 1; i <= defMaxBlockDistance*10; i++ {
   653  		rawblock := protoutil.NewBlock(uint64(i), []byte{})
   654  		b, _ := pb.Marshal(rawblock)
   655  		assert.Error(t, p.s.addPayload(&proto.Payload{
   656  			SeqNum: uint64(i),
   657  			Data:   b,
   658  		}, nonBlocking))
   659  	}
   660  
   661  	// Ensure only blocks 1-4 were passed to the ledger
   662  	close(blocksPassedToLedger)
   663  	i := 1
   664  	for seq := range blocksPassedToLedger {
   665  		assert.Equal(t, uint64(i), seq)
   666  		i++
   667  	}
   668  	assert.Equal(t, 5, i)
   669  
   670  	// Ensure we don't store too many blocks in memory
   671  	sp := p.s
   672  	assert.True(t, sp.payloads.Size() < defMaxBlockDistance)
   673  }
   674  
   675  func TestBlockingEnqueue(t *testing.T) {
   676  	// Scenario: In parallel, get blocks from gossip and from the orderer.
   677  	// The blocks from the orderer we get are X2 times the amount of blocks from gossip.
   678  	// The blocks we get from gossip are random indices, to maximize disruption.
   679  	t.Parallel()
   680  	mc := &mockCommitter{Mock: &mock.Mock{}}
   681  	blocksPassedToLedger := make(chan uint64, 10)
   682  	mc.On("CommitLegacy", mock.Anything).Run(func(arg mock.Arguments) {
   683  		blocksPassedToLedger <- arg.Get(0).(*pcomm.Block).Header.Number
   684  	})
   685  	mc.On("LedgerHeight", mock.Anything).Return(uint64(1), nil)
   686  	mc.On("DoesPvtDataInfoExistInLedger", mock.Anything).Return(false, nil)
   687  	g := &mocks.GossipMock{}
   688  	g.On("Accept", mock.Anything, false).Return(make(<-chan *proto.GossipMessage), nil)
   689  	g.On("Accept", mock.Anything, true).Return(nil, make(chan protoext.ReceivedMessage))
   690  	p := newPeerNode(0, mc, noopPeerIdentityAcceptor)
   691  	defer p.shutdown()
   692  
   693  	numBlocksReceived := 500
   694  	receivedBlockCount := 0
   695  	// Get a block from the orderer every 1ms
   696  	go func() {
   697  		for i := 1; i <= numBlocksReceived; i++ {
   698  			rawblock := protoutil.NewBlock(uint64(i), []byte{})
   699  			b, _ := pb.Marshal(rawblock)
   700  			block := &proto.Payload{
   701  				SeqNum: uint64(i),
   702  				Data:   b,
   703  			}
   704  			p.s.AddPayload(block)
   705  			time.Sleep(time.Millisecond)
   706  		}
   707  	}()
   708  
   709  	// Get a block from gossip every 1ms too
   710  	go func() {
   711  		rand.Seed(time.Now().UnixNano())
   712  		for i := 1; i <= numBlocksReceived/2; i++ {
   713  			blockSeq := rand.Intn(numBlocksReceived)
   714  			rawblock := protoutil.NewBlock(uint64(blockSeq), []byte{})
   715  			b, _ := pb.Marshal(rawblock)
   716  			block := &proto.Payload{
   717  				SeqNum: uint64(blockSeq),
   718  				Data:   b,
   719  			}
   720  			p.s.addPayload(block, nonBlocking)
   721  			time.Sleep(time.Millisecond)
   722  		}
   723  	}()
   724  
   725  	for {
   726  		receivedBlock := <-blocksPassedToLedger
   727  		receivedBlockCount++
   728  		m := &mock.Mock{}
   729  		m.On("LedgerHeight", mock.Anything).Return(receivedBlock, nil)
   730  		m.On("DoesPvtDataInfoExistInLedger", mock.Anything).Return(false, nil)
   731  		m.On("CommitLegacy", mock.Anything).Run(func(arg mock.Arguments) {
   732  			blocksPassedToLedger <- arg.Get(0).(*pcomm.Block).Header.Number
   733  		})
   734  		mc.Lock()
   735  		mc.Mock = m
   736  		mc.Unlock()
   737  		assert.Equal(t, receivedBlock, uint64(receivedBlockCount))
   738  		if int(receivedBlockCount) == numBlocksReceived {
   739  			break
   740  		}
   741  		time.Sleep(time.Millisecond * 10)
   742  	}
   743  }
   744  
   745  func TestHaltChainProcessing(t *testing.T) {
   746  	gossipChannel := func(c chan *proto.GossipMessage) <-chan *proto.GossipMessage {
   747  		return c
   748  	}
   749  	makeBlock := func(seq int) []byte {
   750  		b := &pcomm.Block{
   751  			Header: &pcomm.BlockHeader{
   752  				Number: uint64(seq),
   753  			},
   754  			Data: &pcomm.BlockData{
   755  				Data: [][]byte{},
   756  			},
   757  			Metadata: &pcomm.BlockMetadata{
   758  				Metadata: [][]byte{
   759  					{}, {}, {}, {},
   760  				},
   761  			},
   762  		}
   763  		data, _ := pb.Marshal(b)
   764  		return data
   765  	}
   766  	newBlockMsg := func(i int) *proto.GossipMessage {
   767  		return &proto.GossipMessage{
   768  			Channel: []byte("testchannelid"),
   769  			Content: &proto.GossipMessage_DataMsg{
   770  				DataMsg: &proto.DataMessage{
   771  					Payload: &proto.Payload{
   772  						SeqNum: uint64(i),
   773  						Data:   makeBlock(i),
   774  					},
   775  				},
   776  			},
   777  		}
   778  	}
   779  
   780  	mc := &mockCommitter{Mock: &mock.Mock{}}
   781  	mc.On("CommitLegacy", mock.Anything)
   782  	mc.On("LedgerHeight", mock.Anything).Return(uint64(1), nil)
   783  	g := &mocks.GossipMock{}
   784  	gossipMsgs := make(chan *proto.GossipMessage)
   785  
   786  	g.On("Accept", mock.Anything, false).Return(gossipChannel(gossipMsgs), nil)
   787  	g.On("Accept", mock.Anything, true).Return(nil, make(chan protoext.ReceivedMessage))
   788  	g.On("PeersOfChannel", mock.Anything).Return([]discovery.NetworkMember{})
   789  
   790  	v := &validator.MockValidator{}
   791  	v.On("Validate").Return(&errors2.VSCCExecutionFailureError{
   792  		Err: errors.New("foobar"),
   793  	}).Once()
   794  
   795  	buf := gbytes.NewBuffer()
   796  
   797  	logger := flogging.MustGetLogger(gutil.StateLogger).WithOptions(zap.Hooks(func(entry zapcore.Entry) error {
   798  		buf.Write([]byte(entry.Message))
   799  		buf.Write([]byte("\n"))
   800  		return nil
   801  	}))
   802  	peerNode := newPeerNodeWithGossipWithValidator(logger, 0, mc, noopPeerIdentityAcceptor, g, v)
   803  	defer peerNode.shutdown()
   804  	gossipMsgs <- newBlockMsg(1)
   805  
   806  	gom := gomega.NewGomegaWithT(t)
   807  	gom.Eventually(buf, time.Minute).Should(gbytes.Say("Failed executing VSCC due to foobar"))
   808  	gom.Eventually(buf, time.Minute).Should(gbytes.Say("Aborting chain processing"))
   809  }
   810  
   811  func TestFailures(t *testing.T) {
   812  	t.Parallel()
   813  	mc := &mockCommitter{Mock: &mock.Mock{}}
   814  	mc.On("LedgerHeight", mock.Anything).Return(uint64(0), nil)
   815  	g := &mocks.GossipMock{}
   816  	g.On("Accept", mock.Anything, false).Return(make(<-chan *proto.GossipMessage), nil)
   817  	g.On("Accept", mock.Anything, true).Return(nil, make(chan protoext.ReceivedMessage))
   818  	g.On("PeersOfChannel", mock.Anything).Return([]discovery.NetworkMember{})
   819  	assert.Panics(t, func() {
   820  		newPeerNodeWithGossip(0, mc, noopPeerIdentityAcceptor, g)
   821  	})
   822  	// Reprogram mock
   823  	mc.Mock = &mock.Mock{}
   824  	mc.On("LedgerHeight", mock.Anything).Return(uint64(1), errors.New("Failed accessing ledger"))
   825  	assert.Nil(t, newPeerNodeWithGossip(0, mc, noopPeerIdentityAcceptor, g))
   826  }
   827  
   828  func TestGossipReception(t *testing.T) {
   829  	t.Parallel()
   830  	signalChan := make(chan struct{})
   831  	rawblock := &pcomm.Block{
   832  		Header: &pcomm.BlockHeader{
   833  			Number: uint64(1),
   834  		},
   835  		Data: &pcomm.BlockData{
   836  			Data: [][]byte{},
   837  		},
   838  		Metadata: &pcomm.BlockMetadata{
   839  			Metadata: [][]byte{
   840  				{}, {}, {}, {},
   841  			},
   842  		},
   843  	}
   844  	b, _ := pb.Marshal(rawblock)
   845  
   846  	newMsg := func(channel string) *proto.GossipMessage {
   847  		{
   848  			return &proto.GossipMessage{
   849  				Channel: []byte(channel),
   850  				Content: &proto.GossipMessage_DataMsg{
   851  					DataMsg: &proto.DataMessage{
   852  						Payload: &proto.Payload{
   853  							SeqNum: 1,
   854  							Data:   b,
   855  						},
   856  					},
   857  				},
   858  			}
   859  		}
   860  	}
   861  
   862  	createChan := func(signalChan chan struct{}) <-chan *proto.GossipMessage {
   863  		c := make(chan *proto.GossipMessage)
   864  
   865  		go func(c chan *proto.GossipMessage) {
   866  			// Wait for Accept() to be called
   867  			<-signalChan
   868  			// Simulate a message reception from the gossip component with an invalid channel
   869  			c <- newMsg("AAA")
   870  			// Simulate a message reception from the gossip component
   871  			c <- newMsg("testchannelid")
   872  		}(c)
   873  		return c
   874  	}
   875  
   876  	g := &mocks.GossipMock{}
   877  	rmc := createChan(signalChan)
   878  	g.On("Accept", mock.Anything, false).Return(rmc, nil).Run(func(_ mock.Arguments) {
   879  		signalChan <- struct{}{}
   880  	})
   881  	g.On("Accept", mock.Anything, true).Return(nil, make(chan protoext.ReceivedMessage))
   882  	g.On("PeersOfChannel", mock.Anything).Return([]discovery.NetworkMember{})
   883  	mc := &mockCommitter{Mock: &mock.Mock{}}
   884  	receivedChan := make(chan struct{})
   885  	mc.On("CommitLegacy", mock.Anything).Run(func(arguments mock.Arguments) {
   886  		block := arguments.Get(0).(*pcomm.Block)
   887  		assert.Equal(t, uint64(1), block.Header.Number)
   888  		receivedChan <- struct{}{}
   889  	})
   890  	mc.On("LedgerHeight", mock.Anything).Return(uint64(1), nil)
   891  	mc.On("DoesPvtDataInfoExistInLedger", mock.Anything).Return(false, nil)
   892  	p := newPeerNodeWithGossip(0, mc, noopPeerIdentityAcceptor, g)
   893  	defer p.shutdown()
   894  	select {
   895  	case <-receivedChan:
   896  	case <-time.After(time.Second * 15):
   897  		assert.Fail(t, "Didn't commit a block within a timely manner")
   898  	}
   899  }
   900  
   901  func TestLedgerHeightFromProperties(t *testing.T) {
   902  	// Scenario: For each test, spawn a peer and supply it
   903  	// with a specific mock of PeersOfChannel from peers that
   904  	// either set both metadata properly, or only the properties, or none, or both.
   905  	// Ensure the logic handles all of the 4 possible cases as needed
   906  
   907  	t.Parallel()
   908  	// Returns whether the given networkMember was selected or not
   909  	wasNetworkMemberSelected := func(t *testing.T, networkMember discovery.NetworkMember) bool {
   910  		var wasGivenNetworkMemberSelected int32
   911  		finChan := make(chan struct{})
   912  		g := &mocks.GossipMock{}
   913  		g.On("Send", mock.Anything, mock.Anything).Run(func(arguments mock.Arguments) {
   914  			msg := arguments.Get(0).(*proto.GossipMessage)
   915  			assert.NotNil(t, msg.GetStateRequest())
   916  			peer := arguments.Get(1).([]*comm.RemotePeer)[0]
   917  			if bytes.Equal(networkMember.PKIid, peer.PKIID) {
   918  				atomic.StoreInt32(&wasGivenNetworkMemberSelected, 1)
   919  			}
   920  			finChan <- struct{}{}
   921  		})
   922  		g.On("Accept", mock.Anything, false).Return(make(<-chan *proto.GossipMessage), nil)
   923  		g.On("Accept", mock.Anything, true).Return(nil, make(chan protoext.ReceivedMessage))
   924  		defaultPeer := discovery.NetworkMember{
   925  			InternalEndpoint: "b",
   926  			PKIid:            common.PKIidType("b"),
   927  			Properties: &proto.Properties{
   928  				LedgerHeight: 5,
   929  			},
   930  		}
   931  		g.On("PeersOfChannel", mock.Anything).Return([]discovery.NetworkMember{
   932  			defaultPeer,
   933  			networkMember,
   934  		})
   935  		mc := &mockCommitter{Mock: &mock.Mock{}}
   936  		mc.On("LedgerHeight", mock.Anything).Return(uint64(1), nil)
   937  		p := newPeerNodeWithGossip(0, mc, noopPeerIdentityAcceptor, g)
   938  		defer p.shutdown()
   939  		select {
   940  		case <-time.After(time.Second * 20):
   941  			t.Fatal("Didn't send a request within a timely manner")
   942  		case <-finChan:
   943  		}
   944  		return atomic.LoadInt32(&wasGivenNetworkMemberSelected) == 1
   945  	}
   946  
   947  	peerWithProperties := discovery.NetworkMember{
   948  		PKIid: common.PKIidType("peerWithoutMetadata"),
   949  		Properties: &proto.Properties{
   950  			LedgerHeight: 10,
   951  		},
   952  		InternalEndpoint: "peerWithoutMetadata",
   953  	}
   954  
   955  	peerWithoutProperties := discovery.NetworkMember{
   956  		PKIid:            common.PKIidType("peerWithoutProperties"),
   957  		InternalEndpoint: "peerWithoutProperties",
   958  	}
   959  
   960  	tests := []struct {
   961  		shouldGivenBeSelected bool
   962  		member                discovery.NetworkMember
   963  	}{
   964  		{member: peerWithProperties, shouldGivenBeSelected: true},
   965  		{member: peerWithoutProperties, shouldGivenBeSelected: false},
   966  	}
   967  
   968  	for _, tst := range tests {
   969  		assert.Equal(t, tst.shouldGivenBeSelected, wasNetworkMemberSelected(t, tst.member))
   970  	}
   971  }
   972  
   973  func TestAccessControl(t *testing.T) {
   974  	t.Parallel()
   975  	bootstrapSetSize := 5
   976  	bootstrapSet := make([]*peerNode, 0)
   977  
   978  	authorizedPeersSize := 4
   979  	var listeners []net.Listener
   980  	var endpoints []string
   981  
   982  	for i := 0; i < authorizedPeersSize; i++ {
   983  		ll, err := net.Listen("tcp", "127.0.0.1:0")
   984  		assert.NoError(t, err)
   985  		listeners = append(listeners, ll)
   986  		endpoint := ll.Addr().String()
   987  		endpoints = append(endpoints, endpoint)
   988  	}
   989  
   990  	defer func() {
   991  		for _, ll := range listeners {
   992  			ll.Close()
   993  		}
   994  	}()
   995  
   996  	authorizedPeers := map[string]struct{}{
   997  		endpoints[0]: {},
   998  		endpoints[1]: {},
   999  		endpoints[2]: {},
  1000  		endpoints[3]: {},
  1001  	}
  1002  
  1003  	blockPullPolicy := func(identity api.PeerIdentityType) error {
  1004  		if _, isAuthorized := authorizedPeers[string(identity)]; isAuthorized {
  1005  			return nil
  1006  		}
  1007  		return errors.New("Not authorized")
  1008  	}
  1009  
  1010  	var bootPorts []int
  1011  
  1012  	for i := 0; i < bootstrapSetSize; i++ {
  1013  		commit := newCommitter()
  1014  		bootPeer, bootPort := newBootNode(i, commit, blockPullPolicy)
  1015  		bootstrapSet = append(bootstrapSet, bootPeer)
  1016  		bootPorts = append(bootPorts, bootPort)
  1017  	}
  1018  
  1019  	defer func() {
  1020  		for _, p := range bootstrapSet {
  1021  			p.shutdown()
  1022  		}
  1023  	}()
  1024  
  1025  	msgCount := 5
  1026  
  1027  	for i := 1; i <= msgCount; i++ {
  1028  		rawblock := protoutil.NewBlock(uint64(i), []byte{})
  1029  		if b, err := pb.Marshal(rawblock); err == nil {
  1030  			payload := &proto.Payload{
  1031  				SeqNum: uint64(i),
  1032  				Data:   b,
  1033  			}
  1034  			bootstrapSet[0].s.AddPayload(payload)
  1035  		} else {
  1036  			t.Fail()
  1037  		}
  1038  	}
  1039  
  1040  	standardPeerSetSize := 10
  1041  	peersSet := make([]*peerNode, 0)
  1042  
  1043  	for i := 0; i < standardPeerSetSize; i++ {
  1044  		commit := newCommitter()
  1045  		peersSet = append(peersSet, newPeerNode(bootstrapSetSize+i, commit, blockPullPolicy, bootPorts...))
  1046  	}
  1047  
  1048  	defer func() {
  1049  		for _, p := range peersSet {
  1050  			p.shutdown()
  1051  		}
  1052  	}()
  1053  
  1054  	waitUntilTrueOrTimeout(t, func() bool {
  1055  		for _, p := range peersSet {
  1056  			if len(p.g.PeersOfChannel(common.ChannelID("testchannelid"))) != bootstrapSetSize+standardPeerSetSize-1 {
  1057  				t.Log("Peer discovery has not finished yet")
  1058  				return false
  1059  			}
  1060  		}
  1061  		t.Log("All peer discovered each other!!!")
  1062  		return true
  1063  	}, 30*time.Second)
  1064  
  1065  	t.Log("Waiting for all blocks to arrive.")
  1066  	waitUntilTrueOrTimeout(t, func() bool {
  1067  		t.Log("Trying to see all authorized peers get all blocks, and all non-authorized didn't")
  1068  		for _, p := range peersSet {
  1069  			height, err := p.commit.LedgerHeight()
  1070  			id := fmt.Sprintf("127.0.0.1:%d", p.port)
  1071  			if _, isAuthorized := authorizedPeers[id]; isAuthorized {
  1072  				if height != uint64(msgCount+1) || err != nil {
  1073  					return false
  1074  				}
  1075  			} else {
  1076  				if err == nil && height > 1 {
  1077  					assert.Fail(t, "Peer", id, "got message but isn't authorized! Height:", height)
  1078  				}
  1079  			}
  1080  		}
  1081  		t.Log("All peers have same ledger height!!!")
  1082  		return true
  1083  	}, 60*time.Second)
  1084  }
  1085  
  1086  func TestNewGossipStateProvider_SendingManyMessages(t *testing.T) {
  1087  	t.Parallel()
  1088  	bootstrapSetSize := 5
  1089  	bootstrapSet := make([]*peerNode, 0)
  1090  
  1091  	var bootPorts []int
  1092  
  1093  	for i := 0; i < bootstrapSetSize; i++ {
  1094  		commit := newCommitter()
  1095  		bootPeer, bootPort := newBootNode(i, commit, noopPeerIdentityAcceptor)
  1096  		bootstrapSet = append(bootstrapSet, bootPeer)
  1097  		bootPorts = append(bootPorts, bootPort)
  1098  	}
  1099  
  1100  	defer func() {
  1101  		for _, p := range bootstrapSet {
  1102  			p.shutdown()
  1103  		}
  1104  	}()
  1105  
  1106  	msgCount := 10
  1107  
  1108  	for i := 1; i <= msgCount; i++ {
  1109  		rawblock := protoutil.NewBlock(uint64(i), []byte{})
  1110  		if b, err := pb.Marshal(rawblock); err == nil {
  1111  			payload := &proto.Payload{
  1112  				SeqNum: uint64(i),
  1113  				Data:   b,
  1114  			}
  1115  			bootstrapSet[0].s.AddPayload(payload)
  1116  		} else {
  1117  			t.Fail()
  1118  		}
  1119  	}
  1120  
  1121  	standartPeersSize := 10
  1122  	peersSet := make([]*peerNode, 0)
  1123  
  1124  	for i := 0; i < standartPeersSize; i++ {
  1125  		commit := newCommitter()
  1126  		peersSet = append(peersSet, newPeerNode(bootstrapSetSize+i, commit, noopPeerIdentityAcceptor, bootPorts...))
  1127  	}
  1128  
  1129  	defer func() {
  1130  		for _, p := range peersSet {
  1131  			p.shutdown()
  1132  		}
  1133  	}()
  1134  
  1135  	waitUntilTrueOrTimeout(t, func() bool {
  1136  		for _, p := range peersSet {
  1137  			if len(p.g.PeersOfChannel(common.ChannelID("testchannelid"))) != bootstrapSetSize+standartPeersSize-1 {
  1138  				t.Log("Peer discovery has not finished yet")
  1139  				return false
  1140  			}
  1141  		}
  1142  		t.Log("All peer discovered each other!!!")
  1143  		return true
  1144  	}, 30*time.Second)
  1145  
  1146  	t.Log("Waiting for all blocks to arrive.")
  1147  	waitUntilTrueOrTimeout(t, func() bool {
  1148  		t.Log("Trying to see all peers get all blocks")
  1149  		for _, p := range peersSet {
  1150  			height, err := p.commit.LedgerHeight()
  1151  			if height != uint64(msgCount+1) || err != nil {
  1152  				return false
  1153  			}
  1154  		}
  1155  		t.Log("All peers have same ledger height!!!")
  1156  		return true
  1157  	}, 60*time.Second)
  1158  }
  1159  
  1160  // Start one bootstrap peer and submit defAntiEntropyBatchSize + 5 messages into
  1161  // local ledger, next spawning a new peer waiting for anti-entropy procedure to
  1162  // complete missing blocks. Since state transfer messages now batched, it is expected
  1163  // to see _exactly_ two messages with state transfer response.
  1164  func TestNewGossipStateProvider_BatchingOfStateRequest(t *testing.T) {
  1165  	t.Parallel()
  1166  	bootPeer, bootPort := newBootNode(0, newCommitter(), noopPeerIdentityAcceptor)
  1167  	defer bootPeer.shutdown()
  1168  
  1169  	msgCount := defAntiEntropyBatchSize + 5
  1170  	expectedMessagesCnt := 2
  1171  
  1172  	for i := 1; i <= msgCount; i++ {
  1173  		rawblock := protoutil.NewBlock(uint64(i), []byte{})
  1174  		if b, err := pb.Marshal(rawblock); err == nil {
  1175  			payload := &proto.Payload{
  1176  				SeqNum: uint64(i),
  1177  				Data:   b,
  1178  			}
  1179  			bootPeer.s.AddPayload(payload)
  1180  		} else {
  1181  			t.Fail()
  1182  		}
  1183  	}
  1184  
  1185  	peer := newPeerNode(1, newCommitter(), noopPeerIdentityAcceptor, bootPort)
  1186  	defer peer.shutdown()
  1187  
  1188  	naiveStateMsgPredicate := func(message interface{}) bool {
  1189  		return protoext.IsRemoteStateMessage(message.(protoext.ReceivedMessage).GetGossipMessage().GossipMessage)
  1190  	}
  1191  	_, peerCh := peer.g.Accept(naiveStateMsgPredicate, true)
  1192  
  1193  	wg := sync.WaitGroup{}
  1194  	wg.Add(expectedMessagesCnt)
  1195  
  1196  	// Number of submitted messages is defAntiEntropyBatchSize + 5, therefore
  1197  	// expected number of batches is expectedMessagesCnt = 2. Following go routine
  1198  	// makes sure it receives expected amount of messages and sends signal of success
  1199  	// to continue the test
  1200  	go func() {
  1201  		for count := 0; count < expectedMessagesCnt; count++ {
  1202  			<-peerCh
  1203  			wg.Done()
  1204  		}
  1205  	}()
  1206  
  1207  	// Once we got message which indicate of two batches being received,
  1208  	// making sure messages indeed committed.
  1209  	waitUntilTrueOrTimeout(t, func() bool {
  1210  		if len(peer.g.PeersOfChannel(common.ChannelID("testchannelid"))) != 1 {
  1211  			t.Log("Peer discovery has not finished yet")
  1212  			return false
  1213  		}
  1214  		t.Log("All peer discovered each other!!!")
  1215  		return true
  1216  	}, 30*time.Second)
  1217  
  1218  	// Waits for message which indicates that expected number of message batches received
  1219  	// otherwise timeouts after 2 * defAntiEntropyInterval + 1 seconds
  1220  	wg.Wait()
  1221  
  1222  	t.Log("Waiting for all blocks to arrive.")
  1223  	waitUntilTrueOrTimeout(t, func() bool {
  1224  		t.Log("Trying to see all peers get all blocks")
  1225  		height, err := peer.commit.LedgerHeight()
  1226  		if height != uint64(msgCount+1) || err != nil {
  1227  			return false
  1228  		}
  1229  		t.Log("All peers have same ledger height!!!")
  1230  		return true
  1231  	}, 60*time.Second)
  1232  }
  1233  
  1234  // coordinatorMock mocking structure to capture mock interface for
  1235  // coord to simulate coord flow during the test
  1236  type coordinatorMock struct {
  1237  	committer.Committer
  1238  	mock.Mock
  1239  }
  1240  
  1241  func (mock *coordinatorMock) GetPvtDataAndBlockByNum(seqNum uint64, _ protoutil.SignedData) (*pcomm.Block, gutil.PvtDataCollections, error) {
  1242  	args := mock.Called(seqNum)
  1243  	return args.Get(0).(*pcomm.Block), args.Get(1).(gutil.PvtDataCollections), args.Error(2)
  1244  }
  1245  
  1246  func (mock *coordinatorMock) GetBlockByNum(seqNum uint64) (*pcomm.Block, error) {
  1247  	args := mock.Called(seqNum)
  1248  	return args.Get(0).(*pcomm.Block), args.Error(1)
  1249  }
  1250  
  1251  func (mock *coordinatorMock) StoreBlock(block *pcomm.Block, data gutil.PvtDataCollections) error {
  1252  	args := mock.Called(block, data)
  1253  	return args.Error(1)
  1254  }
  1255  
  1256  func (mock *coordinatorMock) LedgerHeight() (uint64, error) {
  1257  	args := mock.Called()
  1258  	return args.Get(0).(uint64), args.Error(1)
  1259  }
  1260  
  1261  func (mock *coordinatorMock) Close() {
  1262  	mock.Called()
  1263  }
  1264  
  1265  // StorePvtData used to persist private date into transient store
  1266  func (mock *coordinatorMock) StorePvtData(txid string, privData *tspb.TxPvtReadWriteSetWithConfigInfo, blkHeight uint64) error {
  1267  	return mock.Called().Error(0)
  1268  }
  1269  
  1270  type receivedMessageMock struct {
  1271  	mock.Mock
  1272  }
  1273  
  1274  // Ack returns to the sender an acknowledgement for the message
  1275  func (mock *receivedMessageMock) Ack(err error) {
  1276  
  1277  }
  1278  
  1279  func (mock *receivedMessageMock) Respond(msg *proto.GossipMessage) {
  1280  	mock.Called(msg)
  1281  }
  1282  
  1283  func (mock *receivedMessageMock) GetGossipMessage() *protoext.SignedGossipMessage {
  1284  	args := mock.Called()
  1285  	return args.Get(0).(*protoext.SignedGossipMessage)
  1286  }
  1287  
  1288  func (mock *receivedMessageMock) GetSourceEnvelope() *proto.Envelope {
  1289  	args := mock.Called()
  1290  	return args.Get(0).(*proto.Envelope)
  1291  }
  1292  
  1293  func (mock *receivedMessageMock) GetConnectionInfo() *protoext.ConnectionInfo {
  1294  	args := mock.Called()
  1295  	return args.Get(0).(*protoext.ConnectionInfo)
  1296  }
  1297  
  1298  type testData struct {
  1299  	block   *pcomm.Block
  1300  	pvtData gutil.PvtDataCollections
  1301  }
  1302  
  1303  func TestTransferOfPrivateRWSet(t *testing.T) {
  1304  	t.Parallel()
  1305  	chainID := "testChainID"
  1306  
  1307  	// First gossip instance
  1308  	g := &mocks.GossipMock{}
  1309  	coord1 := new(coordinatorMock)
  1310  
  1311  	gossipChannel := make(chan *proto.GossipMessage)
  1312  	commChannel := make(chan protoext.ReceivedMessage)
  1313  
  1314  	gossipChannelFactory := func(ch chan *proto.GossipMessage) <-chan *proto.GossipMessage {
  1315  		return ch
  1316  	}
  1317  
  1318  	g.On("Accept", mock.Anything, false).Return(gossipChannelFactory(gossipChannel), nil)
  1319  	g.On("Accept", mock.Anything, true).Return(nil, commChannel)
  1320  
  1321  	g.On("UpdateChannelMetadata", mock.Anything, mock.Anything)
  1322  	g.On("PeersOfChannel", mock.Anything).Return([]discovery.NetworkMember{})
  1323  	g.On("Close")
  1324  
  1325  	coord1.On("LedgerHeight", mock.Anything).Return(uint64(5), nil)
  1326  
  1327  	var data = map[uint64]*testData{
  1328  		uint64(2): {
  1329  			block: &pcomm.Block{
  1330  				Header: &pcomm.BlockHeader{
  1331  					Number:       2,
  1332  					DataHash:     []byte{0, 1, 1, 1},
  1333  					PreviousHash: []byte{0, 0, 0, 1},
  1334  				},
  1335  				Data: &pcomm.BlockData{
  1336  					Data: [][]byte{{1}, {2}, {3}},
  1337  				},
  1338  			},
  1339  			pvtData: gutil.PvtDataCollections{
  1340  				{
  1341  					SeqInBlock: uint64(0),
  1342  					WriteSet: &rwset.TxPvtReadWriteSet{
  1343  						DataModel: rwset.TxReadWriteSet_KV,
  1344  						NsPvtRwset: []*rwset.NsPvtReadWriteSet{
  1345  							{
  1346  								Namespace: "myCC:v1",
  1347  								CollectionPvtRwset: []*rwset.CollectionPvtReadWriteSet{
  1348  									{
  1349  										CollectionName: "mysecrectCollection",
  1350  										Rwset:          []byte{1, 2, 3, 4, 5},
  1351  									},
  1352  								},
  1353  							},
  1354  						},
  1355  					},
  1356  				},
  1357  			},
  1358  		},
  1359  
  1360  		uint64(3): {
  1361  			block: &pcomm.Block{
  1362  				Header: &pcomm.BlockHeader{
  1363  					Number:       3,
  1364  					DataHash:     []byte{1, 1, 1, 1},
  1365  					PreviousHash: []byte{0, 1, 1, 1},
  1366  				},
  1367  				Data: &pcomm.BlockData{
  1368  					Data: [][]byte{{4}, {5}, {6}},
  1369  				},
  1370  			},
  1371  			pvtData: gutil.PvtDataCollections{
  1372  				{
  1373  					SeqInBlock: uint64(2),
  1374  					WriteSet: &rwset.TxPvtReadWriteSet{
  1375  						DataModel: rwset.TxReadWriteSet_KV,
  1376  						NsPvtRwset: []*rwset.NsPvtReadWriteSet{
  1377  							{
  1378  								Namespace: "otherCC:v1",
  1379  								CollectionPvtRwset: []*rwset.CollectionPvtReadWriteSet{
  1380  									{
  1381  										CollectionName: "topClassified",
  1382  										Rwset:          []byte{0, 0, 0, 4, 2},
  1383  									},
  1384  								},
  1385  							},
  1386  						},
  1387  					},
  1388  				},
  1389  			},
  1390  		},
  1391  	}
  1392  
  1393  	for seqNum, each := range data {
  1394  		coord1.On("GetPvtDataAndBlockByNum", seqNum).Return(each.block, each.pvtData, nil /* no error*/)
  1395  	}
  1396  
  1397  	coord1.On("Close")
  1398  
  1399  	servicesAdapater := &ServicesMediator{GossipAdapter: g, MCSAdapter: &cryptoServiceMock{acceptor: noopPeerIdentityAcceptor}}
  1400  	stateMetrics := metrics.NewGossipMetrics(&disabled.Provider{}).StateMetrics
  1401  	stateConfig := &StateConfig{
  1402  		StateCheckInterval:   DefStateCheckInterval,
  1403  		StateResponseTimeout: DefStateResponseTimeout,
  1404  		StateBatchSize:       DefStateBatchSize,
  1405  		StateMaxRetries:      DefStateMaxRetries,
  1406  		StateBlockBufferSize: DefStateBlockBufferSize,
  1407  		StateChannelSize:     DefStateChannelSize,
  1408  		StateEnabled:         DefStateEnabled,
  1409  	}
  1410  	logger := flogging.MustGetLogger(gutil.StateLogger)
  1411  	st := NewGossipStateProvider(logger, chainID, servicesAdapater, coord1, stateMetrics, blocking, stateConfig)
  1412  	defer st.Stop()
  1413  
  1414  	// Mocked state request message
  1415  	requestMsg := new(receivedMessageMock)
  1416  
  1417  	// Get state request message, blocks [2...3]
  1418  	requestGossipMsg := &proto.GossipMessage{
  1419  		// Copy nonce field from the request, so it will be possible to match response
  1420  		Nonce:   1,
  1421  		Tag:     proto.GossipMessage_CHAN_OR_ORG,
  1422  		Channel: []byte(chainID),
  1423  		Content: &proto.GossipMessage_StateRequest{StateRequest: &proto.RemoteStateRequest{
  1424  			StartSeqNum: 2,
  1425  			EndSeqNum:   3,
  1426  		}},
  1427  	}
  1428  
  1429  	msg, _ := protoext.NoopSign(requestGossipMsg)
  1430  
  1431  	requestMsg.On("GetGossipMessage").Return(msg)
  1432  	requestMsg.On("GetConnectionInfo").Return(&protoext.ConnectionInfo{
  1433  		Auth: &protoext.AuthInfo{},
  1434  	})
  1435  
  1436  	// Channel to send responses back
  1437  	responseChannel := make(chan protoext.ReceivedMessage)
  1438  	defer close(responseChannel)
  1439  
  1440  	requestMsg.On("Respond", mock.Anything).Run(func(args mock.Arguments) {
  1441  		// Get gossip response to respond back on state request
  1442  		response := args.Get(0).(*proto.GossipMessage)
  1443  		// Wrap it up into received response
  1444  		receivedMsg := new(receivedMessageMock)
  1445  		// Create sign response
  1446  		msg, _ := protoext.NoopSign(response)
  1447  		// Mock to respond
  1448  		receivedMsg.On("GetGossipMessage").Return(msg)
  1449  		// Send response
  1450  		responseChannel <- receivedMsg
  1451  	})
  1452  
  1453  	// Send request message via communication channel into state transfer
  1454  	commChannel <- requestMsg
  1455  
  1456  	// State transfer request should result in state response back
  1457  	response := <-responseChannel
  1458  
  1459  	// Start the assertion section
  1460  	stateResponse := response.GetGossipMessage().GetStateResponse()
  1461  
  1462  	assertion := assert.New(t)
  1463  	// Nonce should be equal to Nonce of the request
  1464  	assertion.Equal(response.GetGossipMessage().Nonce, uint64(1))
  1465  	// Payload should not need be nil
  1466  	assertion.NotNil(stateResponse)
  1467  	assertion.NotNil(stateResponse.Payloads)
  1468  	// Exactly two messages expected
  1469  	assertion.Equal(len(stateResponse.Payloads), 2)
  1470  
  1471  	// Assert we have all data and it's same as we expected it
  1472  	for _, each := range stateResponse.Payloads {
  1473  		block := &pcomm.Block{}
  1474  		err := pb.Unmarshal(each.Data, block)
  1475  		assertion.NoError(err)
  1476  
  1477  		assertion.NotNil(block.Header)
  1478  
  1479  		testBlock, ok := data[block.Header.Number]
  1480  		assertion.True(ok)
  1481  
  1482  		for i, d := range testBlock.block.Data.Data {
  1483  			assertion.True(bytes.Equal(d, block.Data.Data[i]))
  1484  		}
  1485  
  1486  		for i, p := range testBlock.pvtData {
  1487  			pvtDataPayload := &proto.PvtDataPayload{}
  1488  			err := pb.Unmarshal(each.PrivateData[i], pvtDataPayload)
  1489  			assertion.NoError(err)
  1490  			pvtRWSet := &rwset.TxPvtReadWriteSet{}
  1491  			err = pb.Unmarshal(pvtDataPayload.Payload, pvtRWSet)
  1492  			assertion.NoError(err)
  1493  			assertion.True(pb.Equal(p.WriteSet, pvtRWSet))
  1494  		}
  1495  	}
  1496  }
  1497  
  1498  type testPeer struct {
  1499  	*mocks.GossipMock
  1500  	id            string
  1501  	gossipChannel chan *proto.GossipMessage
  1502  	commChannel   chan protoext.ReceivedMessage
  1503  	coord         *coordinatorMock
  1504  }
  1505  
  1506  func (t testPeer) Gossip() <-chan *proto.GossipMessage {
  1507  	return t.gossipChannel
  1508  }
  1509  
  1510  func (t testPeer) Comm() chan protoext.ReceivedMessage {
  1511  	return t.commChannel
  1512  }
  1513  
  1514  var peers = map[string]testPeer{
  1515  	"peer1": {
  1516  		id:            "peer1",
  1517  		gossipChannel: make(chan *proto.GossipMessage),
  1518  		commChannel:   make(chan protoext.ReceivedMessage),
  1519  		GossipMock:    &mocks.GossipMock{},
  1520  		coord:         new(coordinatorMock),
  1521  	},
  1522  	"peer2": {
  1523  		id:            "peer2",
  1524  		gossipChannel: make(chan *proto.GossipMessage),
  1525  		commChannel:   make(chan protoext.ReceivedMessage),
  1526  		GossipMock:    &mocks.GossipMock{},
  1527  		coord:         new(coordinatorMock),
  1528  	},
  1529  }
  1530  
  1531  func TestTransferOfPvtDataBetweenPeers(t *testing.T) {
  1532  	/*
  1533  	   This test covers pretty basic scenario, there are two peers: "peer1" and "peer2",
  1534  	   while peer2 missing a few blocks in the ledger therefore asking to replicate those
  1535  	   blocks from the first peers.
  1536  
  1537  	   Test going to check that block from one peer will be replicated into second one and
  1538  	   have identical content.
  1539  	*/
  1540  	t.Parallel()
  1541  	chainID := "testChainID"
  1542  
  1543  	// Initialize peer
  1544  	for _, peer := range peers {
  1545  		peer.On("Accept", mock.Anything, false).Return(peer.Gossip(), nil)
  1546  
  1547  		peer.On("Accept", mock.Anything, true).
  1548  			Return(nil, peer.Comm()).
  1549  			Once().
  1550  			On("Accept", mock.Anything, true).
  1551  			Return(nil, make(chan protoext.ReceivedMessage))
  1552  
  1553  		peer.On("UpdateChannelMetadata", mock.Anything, mock.Anything)
  1554  		peer.coord.On("Close")
  1555  		peer.On("Close")
  1556  	}
  1557  
  1558  	// First peer going to have more advanced ledger
  1559  	peers["peer1"].coord.On("LedgerHeight", mock.Anything).Return(uint64(3), nil)
  1560  
  1561  	// Second peer has a gap of one block, hence it will have to replicate it from previous
  1562  	peers["peer2"].coord.On("LedgerHeight", mock.Anything).Return(uint64(2), nil)
  1563  
  1564  	peers["peer1"].coord.On("DoesPvtDataInfoExistInLedger", mock.Anything).Return(false, nil)
  1565  	peers["peer2"].coord.On("DoesPvtDataInfoExistInLedger", mock.Anything).Return(false, nil)
  1566  
  1567  	peers["peer1"].coord.On("GetPvtDataAndBlockByNum", uint64(2)).Return(&pcomm.Block{
  1568  		Header: &pcomm.BlockHeader{
  1569  			Number:       2,
  1570  			DataHash:     []byte{0, 0, 0, 1},
  1571  			PreviousHash: []byte{0, 1, 1, 1},
  1572  		},
  1573  		Data: &pcomm.BlockData{
  1574  			Data: [][]byte{{4}, {5}, {6}},
  1575  		},
  1576  	}, gutil.PvtDataCollections{&ledger.TxPvtData{
  1577  		SeqInBlock: uint64(1),
  1578  		WriteSet: &rwset.TxPvtReadWriteSet{
  1579  			DataModel: rwset.TxReadWriteSet_KV,
  1580  			NsPvtRwset: []*rwset.NsPvtReadWriteSet{
  1581  				{
  1582  					Namespace: "myCC:v1",
  1583  					CollectionPvtRwset: []*rwset.CollectionPvtReadWriteSet{
  1584  						{
  1585  							CollectionName: "mysecrectCollection",
  1586  							Rwset:          []byte{1, 2, 3, 4, 5},
  1587  						},
  1588  					},
  1589  				},
  1590  			},
  1591  		},
  1592  	}}, nil)
  1593  
  1594  	// Return membership of the peers
  1595  	member2 := discovery.NetworkMember{
  1596  		PKIid:            common.PKIidType([]byte{2}),
  1597  		Endpoint:         "peer2:7051",
  1598  		InternalEndpoint: "peer2:7051",
  1599  		Properties: &proto.Properties{
  1600  			LedgerHeight: 2,
  1601  		},
  1602  	}
  1603  
  1604  	member1 := discovery.NetworkMember{
  1605  		PKIid:            common.PKIidType([]byte{1}),
  1606  		Endpoint:         "peer1:7051",
  1607  		InternalEndpoint: "peer1:7051",
  1608  		Properties: &proto.Properties{
  1609  			LedgerHeight: 3,
  1610  		},
  1611  	}
  1612  
  1613  	peers["peer1"].On("PeersOfChannel", mock.Anything).Return([]discovery.NetworkMember{member2})
  1614  	peers["peer2"].On("PeersOfChannel", mock.Anything).Return([]discovery.NetworkMember{member1})
  1615  
  1616  	peers["peer2"].On("Send", mock.Anything, mock.Anything).Run(func(args mock.Arguments) {
  1617  		request := args.Get(0).(*proto.GossipMessage)
  1618  		requestMsg := new(receivedMessageMock)
  1619  		msg, _ := protoext.NoopSign(request)
  1620  		requestMsg.On("GetGossipMessage").Return(msg)
  1621  		requestMsg.On("GetConnectionInfo").Return(&protoext.ConnectionInfo{
  1622  			Auth: &protoext.AuthInfo{},
  1623  		})
  1624  
  1625  		requestMsg.On("Respond", mock.Anything).Run(func(args mock.Arguments) {
  1626  			response := args.Get(0).(*proto.GossipMessage)
  1627  			receivedMsg := new(receivedMessageMock)
  1628  			msg, _ := protoext.NoopSign(response)
  1629  			receivedMsg.On("GetGossipMessage").Return(msg)
  1630  			// Send response back to the peer
  1631  			peers["peer2"].commChannel <- receivedMsg
  1632  		})
  1633  
  1634  		peers["peer1"].commChannel <- requestMsg
  1635  	})
  1636  
  1637  	wg := sync.WaitGroup{}
  1638  	wg.Add(1)
  1639  	peers["peer2"].coord.On("StoreBlock", mock.Anything, mock.Anything).Run(func(args mock.Arguments) {
  1640  		wg.Done() // Done once second peer hits commit of the block
  1641  	}).Return([]string{}, nil) // No pvt data to complete and no error
  1642  
  1643  	cryptoService := &cryptoServiceMock{acceptor: noopPeerIdentityAcceptor}
  1644  
  1645  	stateMetrics := metrics.NewGossipMetrics(&disabled.Provider{}).StateMetrics
  1646  
  1647  	mediator := &ServicesMediator{GossipAdapter: peers["peer1"], MCSAdapter: cryptoService}
  1648  	stateConfig := &StateConfig{
  1649  		StateCheckInterval:   DefStateCheckInterval,
  1650  		StateResponseTimeout: DefStateResponseTimeout,
  1651  		StateBatchSize:       DefStateBatchSize,
  1652  		StateMaxRetries:      DefStateMaxRetries,
  1653  		StateBlockBufferSize: DefStateBlockBufferSize,
  1654  		StateChannelSize:     DefStateChannelSize,
  1655  		StateEnabled:         DefStateEnabled,
  1656  	}
  1657  	logger := flogging.MustGetLogger(gutil.StateLogger)
  1658  	peer1State := NewGossipStateProvider(logger, chainID, mediator, peers["peer1"].coord, stateMetrics, blocking, stateConfig)
  1659  	defer peer1State.Stop()
  1660  
  1661  	mediator = &ServicesMediator{GossipAdapter: peers["peer2"], MCSAdapter: cryptoService}
  1662  	logger = flogging.MustGetLogger(gutil.StateLogger)
  1663  	peer2State := NewGossipStateProvider(logger, chainID, mediator, peers["peer2"].coord, stateMetrics, blocking, stateConfig)
  1664  	defer peer2State.Stop()
  1665  
  1666  	// Make sure state was replicated
  1667  	done := make(chan struct{})
  1668  	go func() {
  1669  		wg.Wait()
  1670  		done <- struct{}{}
  1671  	}()
  1672  
  1673  	select {
  1674  	case <-done:
  1675  		break
  1676  	case <-time.After(30 * time.Second):
  1677  		t.Fail()
  1678  	}
  1679  }
  1680  
  1681  func TestStateRequestValidator(t *testing.T) {
  1682  	validator := &stateRequestValidator{}
  1683  	err := validator.validate(&proto.RemoteStateRequest{
  1684  		StartSeqNum: 10,
  1685  		EndSeqNum:   5,
  1686  	}, defAntiEntropyBatchSize)
  1687  	assert.Contains(t, err.Error(), "Invalid sequence interval [10...5).")
  1688  	assert.Error(t, err)
  1689  
  1690  	err = validator.validate(&proto.RemoteStateRequest{
  1691  		StartSeqNum: 10,
  1692  		EndSeqNum:   30,
  1693  	}, defAntiEntropyBatchSize)
  1694  	assert.Contains(t, err.Error(), "Requesting blocks range [10-30) greater than configured")
  1695  	assert.Error(t, err)
  1696  
  1697  	err = validator.validate(&proto.RemoteStateRequest{
  1698  		StartSeqNum: 10,
  1699  		EndSeqNum:   20,
  1700  	}, defAntiEntropyBatchSize)
  1701  	assert.NoError(t, err)
  1702  }
  1703  
  1704  func waitUntilTrueOrTimeout(t *testing.T, predicate func() bool, timeout time.Duration) {
  1705  	ch := make(chan struct{})
  1706  	t.Log("Started to spin off, until predicate will be satisfied.")
  1707  
  1708  	go func() {
  1709  		t := time.NewTicker(time.Second)
  1710  		for !predicate() {
  1711  			select {
  1712  			case <-ch:
  1713  				t.Stop()
  1714  				return
  1715  			case <-t.C:
  1716  			}
  1717  		}
  1718  		t.Stop()
  1719  		close(ch)
  1720  	}()
  1721  
  1722  	select {
  1723  	case <-ch:
  1724  		t.Log("Done.")
  1725  		break
  1726  	case <-time.After(timeout):
  1727  		t.Fatal("Timeout has expired")
  1728  		close(ch)
  1729  		break
  1730  	}
  1731  	t.Log("Stop waiting until timeout or true")
  1732  }