github.com/whtcorpsinc/MilevaDB-Prod@v0.0.0-20211104133533-f57f4be3b597/interlock/split_test.go (about)

     1  // Copyright 2020 WHTCORPS INC, Inc.
     2  //
     3  // Licensed under the Apache License, Version 2.0 (the "License");
     4  // you may not use this file except in compliance with the License.
     5  // You may obtain a copy of the License at
     6  //
     7  //     http://www.apache.org/licenses/LICENSE-2.0
     8  //
     9  // Unless required by applicable law or agreed to in writing, software
    10  // distributed under the License is distributed on an "AS IS" BASIS,
    11  // See the License for the specific language governing permissions and
    12  // limitations under the License.
    13  
    14  package interlock
    15  
    16  import (
    17  	"bytes"
    18  	"encoding/binary"
    19  	"math"
    20  	"math/rand"
    21  	"sort"
    22  	"time"
    23  
    24  	"github.com/whtcorpsinc/BerolinaSQL/allegrosql"
    25  	"github.com/whtcorpsinc/BerolinaSQL/perceptron"
    26  	. "github.com/whtcorpsinc/check"
    27  	"github.com/whtcorpsinc/milevadb/blockcodec"
    28  	"github.com/whtcorpsinc/milevadb/causet/blocks"
    29  	"github.com/whtcorpsinc/milevadb/causet/embedded"
    30  	"github.com/whtcorpsinc/milevadb/ekv"
    31  	"github.com/whtcorpsinc/milevadb/memex"
    32  	"github.com/whtcorpsinc/milevadb/soliton/mock"
    33  	"github.com/whtcorpsinc/milevadb/stochastikctx/stmtctx"
    34  	"github.com/whtcorpsinc/milevadb/types"
    35  )
    36  
    37  var _ = Suite(&testSplitIndex{})
    38  
    39  type testSplitIndex struct {
    40  }
    41  
    42  func (s *testSplitIndex) SetUpSuite(c *C) {
    43  }
    44  
    45  func (s *testSplitIndex) TearDownSuite(c *C) {
    46  }
    47  
    48  func (s *testSplitIndex) TestLongestCommonPrefixLen(c *C) {
    49  	cases := []struct {
    50  		s1 string
    51  		s2 string
    52  		l  int
    53  	}{
    54  		{"", "", 0},
    55  		{"", "a", 0},
    56  		{"a", "", 0},
    57  		{"a", "a", 1},
    58  		{"ab", "a", 1},
    59  		{"a", "ab", 1},
    60  		{"b", "ab", 0},
    61  		{"ba", "ab", 0},
    62  	}
    63  
    64  	for _, ca := range cases {
    65  		re := longestCommonPrefixLen([]byte(ca.s1), []byte(ca.s2))
    66  		c.Assert(re, Equals, ca.l)
    67  	}
    68  }
    69  
    70  func (s *testSplitIndex) TestgetStepValue(c *C) {
    71  	cases := []struct {
    72  		lower []byte
    73  		upper []byte
    74  		l     int
    75  		v     uint64
    76  	}{
    77  		{[]byte{}, []byte{}, 0, math.MaxUint64},
    78  		{[]byte{0}, []byte{128}, 0, binary.BigEndian.Uint64([]byte{128, 255, 255, 255, 255, 255, 255, 255})},
    79  		{[]byte{'a'}, []byte{'z'}, 0, binary.BigEndian.Uint64([]byte{'z' - 'a', 255, 255, 255, 255, 255, 255, 255})},
    80  		{[]byte("abc"), []byte{'z'}, 0, binary.BigEndian.Uint64([]byte{'z' - 'a', 255 - 'b', 255 - 'c', 255, 255, 255, 255, 255})},
    81  		{[]byte("abc"), []byte("xyz"), 0, binary.BigEndian.Uint64([]byte{'x' - 'a', 'y' - 'b', 'z' - 'c', 255, 255, 255, 255, 255})},
    82  		{[]byte("abc"), []byte("axyz"), 1, binary.BigEndian.Uint64([]byte{'x' - 'b', 'y' - 'c', 'z', 255, 255, 255, 255, 255})},
    83  		{[]byte("abc0123456"), []byte("xyz01234"), 0, binary.BigEndian.Uint64([]byte{'x' - 'a', 'y' - 'b', 'z' - 'c', 0, 0, 0, 0, 0})},
    84  	}
    85  
    86  	for _, ca := range cases {
    87  		l := longestCommonPrefixLen(ca.lower, ca.upper)
    88  		c.Assert(l, Equals, ca.l)
    89  		v0 := getStepValue(ca.lower[l:], ca.upper[l:], 1)
    90  		c.Assert(v0, Equals, ca.v)
    91  	}
    92  }
    93  
    94  func (s *testSplitIndex) TestSplitIndex(c *C) {
    95  	tbInfo := &perceptron.BlockInfo{
    96  		Name: perceptron.NewCIStr("t1"),
    97  		ID:   rand.Int63(),
    98  		DeferredCausets: []*perceptron.DeferredCausetInfo{
    99  			{
   100  				Name:         perceptron.NewCIStr("c0"),
   101  				ID:           1,
   102  				Offset:       1,
   103  				DefaultValue: 0,
   104  				State:        perceptron.StatePublic,
   105  				FieldType:    *types.NewFieldType(allegrosql.TypeLong),
   106  			},
   107  		},
   108  	}
   109  	idxDefCauss := []*perceptron.IndexDeferredCauset{{Name: tbInfo.DeferredCausets[0].Name, Offset: 0, Length: types.UnspecifiedLength}}
   110  	idxInfo := &perceptron.IndexInfo{
   111  		ID:              2,
   112  		Name:            perceptron.NewCIStr("idx1"),
   113  		Block:           perceptron.NewCIStr("t1"),
   114  		DeferredCausets: idxDefCauss,
   115  		State:           perceptron.StatePublic,
   116  	}
   117  	firstIdxInfo0 := idxInfo.Clone()
   118  	firstIdxInfo0.ID = 1
   119  	firstIdxInfo0.Name = perceptron.NewCIStr("idx")
   120  	tbInfo.Indices = []*perceptron.IndexInfo{firstIdxInfo0, idxInfo}
   121  
   122  	// Test for int index.
   123  	// range is 0 ~ 100, and split into 10 region.
   124  	// So 10 regions range is like below, left close right open interval:
   125  	// region1: [-inf ~ 10)
   126  	// region2: [10 ~ 20)
   127  	// region3: [20 ~ 30)
   128  	// region4: [30 ~ 40)
   129  	// region5: [40 ~ 50)
   130  	// region6: [50 ~ 60)
   131  	// region7: [60 ~ 70)
   132  	// region8: [70 ~ 80)
   133  	// region9: [80 ~ 90)
   134  	// region10: [90 ~ +inf)
   135  	ctx := mock.NewContext()
   136  	e := &SplitIndexRegionInterDirc{
   137  		baseInterlockingDirectorate: newBaseInterlockingDirectorate(ctx, nil, 0),
   138  		blockInfo:                   tbInfo,
   139  		indexInfo:                   idxInfo,
   140  		lower:                       []types.Causet{types.NewCauset(0)},
   141  		upper:                       []types.Causet{types.NewCauset(100)},
   142  		num:                         10,
   143  	}
   144  	valueList, err := e.getSplitIdxKeys()
   145  	sort.Slice(valueList, func(i, j int) bool { return bytes.Compare(valueList[i], valueList[j]) < 0 })
   146  	c.Assert(err, IsNil)
   147  	c.Assert(len(valueList), Equals, e.num+1)
   148  
   149  	cases := []struct {
   150  		value        int
   151  		lessEqualIdx int
   152  	}{
   153  		{-1, 0},
   154  		{0, 0},
   155  		{1, 0},
   156  		{10, 1},
   157  		{11, 1},
   158  		{20, 2},
   159  		{21, 2},
   160  		{31, 3},
   161  		{41, 4},
   162  		{51, 5},
   163  		{61, 6},
   164  		{71, 7},
   165  		{81, 8},
   166  		{91, 9},
   167  		{100, 9},
   168  		{1000, 9},
   169  	}
   170  
   171  	index := blocks.NewIndex(tbInfo.ID, tbInfo, idxInfo)
   172  	for _, ca := range cases {
   173  		// test for minInt64 handle
   174  		idxValue, _, err := index.GenIndexKey(ctx.GetStochastikVars().StmtCtx, []types.Causet{types.NewCauset(ca.value)}, ekv.IntHandle(math.MinInt64), nil)
   175  		c.Assert(err, IsNil)
   176  		idx := searchLessEqualIdx(valueList, idxValue)
   177  		c.Assert(idx, Equals, ca.lessEqualIdx, Commentf("%#v", ca))
   178  
   179  		// Test for max int64 handle.
   180  		idxValue, _, err = index.GenIndexKey(ctx.GetStochastikVars().StmtCtx, []types.Causet{types.NewCauset(ca.value)}, ekv.IntHandle(math.MaxInt64), nil)
   181  		c.Assert(err, IsNil)
   182  		idx = searchLessEqualIdx(valueList, idxValue)
   183  		c.Assert(idx, Equals, ca.lessEqualIdx, Commentf("%#v", ca))
   184  	}
   185  	// Test for varchar index.
   186  	// range is a ~ z, and split into 26 region.
   187  	// So 26 regions range is like below:
   188  	// region1: [-inf ~ b)
   189  	// region2: [b ~ c)
   190  	// .
   191  	// .
   192  	// .
   193  	// region26: [y ~ +inf)
   194  	e.lower = []types.Causet{types.NewCauset("a")}
   195  	e.upper = []types.Causet{types.NewCauset("z")}
   196  	e.num = 26
   197  	// change index defCausumn type to varchar
   198  	tbInfo.DeferredCausets[0].FieldType = *types.NewFieldType(allegrosql.TypeVarchar)
   199  
   200  	valueList, err = e.getSplitIdxKeys()
   201  	sort.Slice(valueList, func(i, j int) bool { return bytes.Compare(valueList[i], valueList[j]) < 0 })
   202  	c.Assert(err, IsNil)
   203  	c.Assert(len(valueList), Equals, e.num+1)
   204  
   205  	cases2 := []struct {
   206  		value        string
   207  		lessEqualIdx int
   208  	}{
   209  		{"", 0},
   210  		{"a", 0},
   211  		{"abcde", 0},
   212  		{"b", 1},
   213  		{"bzzzz", 1},
   214  		{"c", 2},
   215  		{"czzzz", 2},
   216  		{"z", 25},
   217  		{"zabcd", 25},
   218  	}
   219  
   220  	for _, ca := range cases2 {
   221  		// test for minInt64 handle
   222  		idxValue, _, err := index.GenIndexKey(ctx.GetStochastikVars().StmtCtx, []types.Causet{types.NewCauset(ca.value)}, ekv.IntHandle(math.MinInt64), nil)
   223  		c.Assert(err, IsNil)
   224  		idx := searchLessEqualIdx(valueList, idxValue)
   225  		c.Assert(idx, Equals, ca.lessEqualIdx, Commentf("%#v", ca))
   226  
   227  		// Test for max int64 handle.
   228  		idxValue, _, err = index.GenIndexKey(ctx.GetStochastikVars().StmtCtx, []types.Causet{types.NewCauset(ca.value)}, ekv.IntHandle(math.MaxInt64), nil)
   229  		c.Assert(err, IsNil)
   230  		idx = searchLessEqualIdx(valueList, idxValue)
   231  		c.Assert(idx, Equals, ca.lessEqualIdx, Commentf("%#v", ca))
   232  	}
   233  
   234  	// Test for timestamp index.
   235  	// range is 2010-01-01 00:00:00 ~ 2020-01-01 00:00:00, and split into 10 region.
   236  	// So 10 regions range is like below:
   237  	// region1: [-inf					~ 2011-01-01 00:00:00)
   238  	// region2: [2011-01-01 00:00:00 	~ 2012-01-01 00:00:00)
   239  	// .
   240  	// .
   241  	// .
   242  	// region10: [2020-01-01 00:00:00 	~ +inf)
   243  	lowerTime := types.NewTime(types.FromDate(2010, 1, 1, 0, 0, 0, 0), allegrosql.TypeTimestamp, types.DefaultFsp)
   244  	upperTime := types.NewTime(types.FromDate(2020, 1, 1, 0, 0, 0, 0), allegrosql.TypeTimestamp, types.DefaultFsp)
   245  	e.lower = []types.Causet{types.NewCauset(lowerTime)}
   246  	e.upper = []types.Causet{types.NewCauset(upperTime)}
   247  	e.num = 10
   248  
   249  	// change index defCausumn type to timestamp
   250  	tbInfo.DeferredCausets[0].FieldType = *types.NewFieldType(allegrosql.TypeTimestamp)
   251  
   252  	valueList, err = e.getSplitIdxKeys()
   253  	sort.Slice(valueList, func(i, j int) bool { return bytes.Compare(valueList[i], valueList[j]) < 0 })
   254  	c.Assert(err, IsNil)
   255  	c.Assert(len(valueList), Equals, e.num+1)
   256  
   257  	cases3 := []struct {
   258  		value        types.CoreTime
   259  		lessEqualIdx int
   260  	}{
   261  		{types.FromDate(2009, 11, 20, 12, 50, 59, 0), 0},
   262  		{types.FromDate(2010, 1, 1, 0, 0, 0, 0), 0},
   263  		{types.FromDate(2011, 12, 31, 23, 59, 59, 0), 1},
   264  		{types.FromDate(2011, 2, 1, 0, 0, 0, 0), 1},
   265  		{types.FromDate(2012, 3, 1, 0, 0, 0, 0), 2},
   266  		{types.FromDate(2020, 4, 1, 0, 0, 0, 0), 3},
   267  		{types.FromDate(2020, 5, 1, 0, 0, 0, 0), 4},
   268  		{types.FromDate(2020, 6, 1, 0, 0, 0, 0), 5},
   269  		{types.FromDate(2020, 8, 1, 0, 0, 0, 0), 6},
   270  		{types.FromDate(2020, 9, 1, 0, 0, 0, 0), 7},
   271  		{types.FromDate(2020, 10, 1, 0, 0, 0, 0), 8},
   272  		{types.FromDate(2020, 11, 1, 0, 0, 0, 0), 9},
   273  		{types.FromDate(2020, 12, 1, 0, 0, 0, 0), 9},
   274  		{types.FromDate(2030, 12, 1, 0, 0, 0, 0), 9},
   275  	}
   276  
   277  	for _, ca := range cases3 {
   278  		value := types.NewTime(ca.value, allegrosql.TypeTimestamp, types.DefaultFsp)
   279  		// test for min int64 handle
   280  		idxValue, _, err := index.GenIndexKey(ctx.GetStochastikVars().StmtCtx, []types.Causet{types.NewCauset(value)}, ekv.IntHandle(math.MinInt64), nil)
   281  		c.Assert(err, IsNil)
   282  		idx := searchLessEqualIdx(valueList, idxValue)
   283  		c.Assert(idx, Equals, ca.lessEqualIdx, Commentf("%#v", ca))
   284  
   285  		// Test for max int64 handle.
   286  		idxValue, _, err = index.GenIndexKey(ctx.GetStochastikVars().StmtCtx, []types.Causet{types.NewCauset(value)}, ekv.IntHandle(math.MaxInt64), nil)
   287  		c.Assert(err, IsNil)
   288  		idx = searchLessEqualIdx(valueList, idxValue)
   289  		c.Assert(idx, Equals, ca.lessEqualIdx, Commentf("%#v", ca))
   290  	}
   291  }
   292  
   293  func (s *testSplitIndex) TestSplitBlock(c *C) {
   294  	tbInfo := &perceptron.BlockInfo{
   295  		Name: perceptron.NewCIStr("t1"),
   296  		ID:   rand.Int63(),
   297  		DeferredCausets: []*perceptron.DeferredCausetInfo{
   298  			{
   299  				Name:         perceptron.NewCIStr("c0"),
   300  				ID:           1,
   301  				Offset:       1,
   302  				DefaultValue: 0,
   303  				State:        perceptron.StatePublic,
   304  				FieldType:    *types.NewFieldType(allegrosql.TypeLong),
   305  			},
   306  		},
   307  	}
   308  	defer func(originValue int64) {
   309  		minRegionStepValue = originValue
   310  	}(minRegionStepValue)
   311  	minRegionStepValue = 10
   312  	// range is 0 ~ 100, and split into 10 region.
   313  	// So 10 regions range is like below:
   314  	// region1: [-inf ~ 10)
   315  	// region2: [10 ~ 20)
   316  	// region3: [20 ~ 30)
   317  	// region4: [30 ~ 40)
   318  	// region5: [40 ~ 50)
   319  	// region6: [50 ~ 60)
   320  	// region7: [60 ~ 70)
   321  	// region8: [70 ~ 80)
   322  	// region9: [80 ~ 90 )
   323  	// region10: [90 ~ +inf)
   324  	ctx := mock.NewContext()
   325  	e := &SplitBlockRegionInterDirc{
   326  		baseInterlockingDirectorate: newBaseInterlockingDirectorate(ctx, nil, 0),
   327  		blockInfo:                   tbInfo,
   328  		handleDefCauss:              embedded.NewIntHandleDefCauss(&memex.DeferredCauset{RetType: types.NewFieldType(allegrosql.TypeLonglong)}),
   329  		lower:                       []types.Causet{types.NewCauset(0)},
   330  		upper:                       []types.Causet{types.NewCauset(100)},
   331  		num:                         10,
   332  	}
   333  	valueList, err := e.getSplitBlockKeys()
   334  	c.Assert(err, IsNil)
   335  	c.Assert(len(valueList), Equals, e.num-1)
   336  
   337  	cases := []struct {
   338  		value        int
   339  		lessEqualIdx int
   340  	}{
   341  		{-1, -1},
   342  		{0, -1},
   343  		{1, -1},
   344  		{10, 0},
   345  		{11, 0},
   346  		{20, 1},
   347  		{21, 1},
   348  		{31, 2},
   349  		{41, 3},
   350  		{51, 4},
   351  		{61, 5},
   352  		{71, 6},
   353  		{81, 7},
   354  		{91, 8},
   355  		{100, 8},
   356  		{1000, 8},
   357  	}
   358  
   359  	recordPrefix := blockcodec.GenBlockRecordPrefix(e.blockInfo.ID)
   360  	for _, ca := range cases {
   361  		// test for minInt64 handle
   362  		key := blockcodec.EncodeRecordKey(recordPrefix, ekv.IntHandle(ca.value))
   363  		c.Assert(err, IsNil)
   364  		idx := searchLessEqualIdx(valueList, key)
   365  		c.Assert(idx, Equals, ca.lessEqualIdx, Commentf("%#v", ca))
   366  	}
   367  }
   368  
   369  func (s *testSplitIndex) TestClusterIndexSplitBlock(c *C) {
   370  	tbInfo := &perceptron.BlockInfo{
   371  		Name:           perceptron.NewCIStr("t"),
   372  		ID:             1,
   373  		IsCommonHandle: true,
   374  		Indices: []*perceptron.IndexInfo{
   375  			{
   376  				ID:      1,
   377  				Primary: true,
   378  				State:   perceptron.StatePublic,
   379  				DeferredCausets: []*perceptron.IndexDeferredCauset{
   380  					{Offset: 1},
   381  					{Offset: 2},
   382  				},
   383  			},
   384  		},
   385  		DeferredCausets: []*perceptron.DeferredCausetInfo{
   386  			{
   387  				Name:      perceptron.NewCIStr("c0"),
   388  				ID:        1,
   389  				Offset:    0,
   390  				State:     perceptron.StatePublic,
   391  				FieldType: *types.NewFieldType(allegrosql.TypeDouble),
   392  			},
   393  			{
   394  				Name:      perceptron.NewCIStr("c1"),
   395  				ID:        2,
   396  				Offset:    1,
   397  				State:     perceptron.StatePublic,
   398  				FieldType: *types.NewFieldType(allegrosql.TypeLonglong),
   399  			},
   400  			{
   401  				Name:      perceptron.NewCIStr("c2"),
   402  				ID:        3,
   403  				Offset:    2,
   404  				State:     perceptron.StatePublic,
   405  				FieldType: *types.NewFieldType(allegrosql.TypeLonglong),
   406  			},
   407  		},
   408  	}
   409  	defer func(originValue int64) {
   410  		minRegionStepValue = originValue
   411  	}(minRegionStepValue)
   412  	minRegionStepValue = 3
   413  	ctx := mock.NewContext()
   414  	sc := &stmtctx.StatementContext{TimeZone: time.Local}
   415  	e := &SplitBlockRegionInterDirc{
   416  		baseInterlockingDirectorate: newBaseInterlockingDirectorate(ctx, nil, 0),
   417  		blockInfo:                   tbInfo,
   418  		handleDefCauss:              buildHandleDefCaussForSplit(sc, tbInfo),
   419  		lower:                       types.MakeCausets(1, 0),
   420  		upper:                       types.MakeCausets(1, 100),
   421  		num:                         10,
   422  	}
   423  	valueList, err := e.getSplitBlockKeys()
   424  	c.Assert(err, IsNil)
   425  	c.Assert(len(valueList), Equals, e.num-1)
   426  
   427  	cases := []struct {
   428  		value        []types.Causet
   429  		lessEqualIdx int
   430  	}{
   431  		// For lower-bound and upper-bound, because 0 and 100 are padding with 7 zeros,
   432  		// the split points are not (i * 10) but approximation.
   433  		{types.MakeCausets(1, -1), -1},
   434  		{types.MakeCausets(1, 0), -1},
   435  		{types.MakeCausets(1, 10), -1},
   436  		{types.MakeCausets(1, 11), 0},
   437  		{types.MakeCausets(1, 20), 0},
   438  		{types.MakeCausets(1, 21), 1},
   439  
   440  		{types.MakeCausets(1, 31), 2},
   441  		{types.MakeCausets(1, 41), 3},
   442  		{types.MakeCausets(1, 51), 4},
   443  		{types.MakeCausets(1, 61), 5},
   444  		{types.MakeCausets(1, 71), 6},
   445  		{types.MakeCausets(1, 81), 7},
   446  		{types.MakeCausets(1, 91), 8},
   447  		{types.MakeCausets(1, 100), 8},
   448  		{types.MakeCausets(1, 101), 8},
   449  	}
   450  
   451  	recordPrefix := blockcodec.GenBlockRecordPrefix(e.blockInfo.ID)
   452  	for _, ca := range cases {
   453  		h, err := e.handleDefCauss.BuildHandleByCausets(ca.value)
   454  		c.Assert(err, IsNil)
   455  		key := blockcodec.EncodeRecordKey(recordPrefix, h)
   456  		c.Assert(err, IsNil)
   457  		idx := searchLessEqualIdx(valueList, key)
   458  		c.Assert(idx, Equals, ca.lessEqualIdx, Commentf("%#v", ca))
   459  	}
   460  }
   461  
   462  func searchLessEqualIdx(valueList [][]byte, value []byte) int {
   463  	idx := -1
   464  	for i, v := range valueList {
   465  		if bytes.Compare(value, v) >= 0 {
   466  			idx = i
   467  			continue
   468  		}
   469  		break
   470  	}
   471  	return idx
   472  }