github.com/parquet-go/parquet-go@v0.20.0/dictionary_test.go (about)

     1  package parquet_test
     2  
     3  import (
     4  	"bytes"
     5  	"fmt"
     6  	"math/rand"
     7  	"testing"
     8  	"time"
     9  
    10  	"github.com/parquet-go/parquet-go"
    11  )
    12  
    13  var dictionaryTypes = [...]parquet.Type{
    14  	parquet.BooleanType,
    15  	parquet.Int32Type,
    16  	parquet.Int64Type,
    17  	parquet.Int96Type,
    18  	parquet.FloatType,
    19  	parquet.DoubleType,
    20  	parquet.ByteArrayType,
    21  	parquet.FixedLenByteArrayType(10),
    22  	parquet.FixedLenByteArrayType(16),
    23  	parquet.Uint(32).Type(),
    24  	parquet.Uint(64).Type(),
    25  }
    26  
    27  func TestDictionary(t *testing.T) {
    28  	for _, typ := range dictionaryTypes {
    29  		for _, numValues := range []int{1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 17, 1e2, 1e3, 1e4} {
    30  			t.Run(fmt.Sprintf("%s/N=%d", typ, numValues), func(t *testing.T) {
    31  				testDictionary(t, typ, numValues)
    32  			})
    33  		}
    34  	}
    35  }
    36  
    37  func testDictionary(t *testing.T, typ parquet.Type, numValues int) {
    38  	const columnIndex = 1
    39  
    40  	dict := typ.NewDictionary(columnIndex, 0, typ.NewValues(nil, nil))
    41  	values := make([]parquet.Value, numValues)
    42  	indexes := make([]int32, numValues)
    43  	lookups := make([]parquet.Value, numValues)
    44  
    45  	f := randValueFuncOf(typ)
    46  	r := rand.New(rand.NewSource(int64(numValues)))
    47  
    48  	for i := range values {
    49  		values[i] = f(r)
    50  		values[i] = values[i].Level(0, 0, columnIndex)
    51  	}
    52  
    53  	mapping := make(map[int32]parquet.Value, numValues)
    54  
    55  	for i := 0; i < numValues; {
    56  		j := i + ((numValues-i)/2 + 1)
    57  		if j > numValues {
    58  			j = numValues
    59  		}
    60  
    61  		dict.Insert(indexes[i:j], values[i:j])
    62  
    63  		for k, v := range values[i:j] {
    64  			mapping[indexes[i+k]] = v
    65  		}
    66  
    67  		for _, index := range indexes[i:j] {
    68  			if index < 0 || index >= int32(dict.Len()) {
    69  				t.Fatalf("index out of bounds: %d", index)
    70  			}
    71  		}
    72  
    73  		// second insert is a no-op since all the values are already in the dictionary
    74  		lastDictLen := dict.Len()
    75  		dict.Insert(indexes[i:j], values[i:j])
    76  
    77  		if dict.Len() != lastDictLen {
    78  			for k, index := range indexes[i:j] {
    79  				if index >= int32(len(mapping)) {
    80  					t.Log(values[i+k])
    81  				}
    82  			}
    83  
    84  			t.Fatalf("%d values were inserted on the second pass", dict.Len()-len(mapping))
    85  		}
    86  
    87  		r.Shuffle(j-i, func(a, b int) {
    88  			indexes[a+i], indexes[b+i] = indexes[b+i], indexes[a+i]
    89  		})
    90  
    91  		dict.Lookup(indexes[i:j], lookups[i:j])
    92  
    93  		for lookupIndex, valueIndex := range indexes[i:j] {
    94  			want := mapping[valueIndex]
    95  			got := lookups[lookupIndex+i]
    96  
    97  			if !parquet.DeepEqual(want, got) {
    98  				t.Fatalf("wrong value looked up at index %d: want=%#v got=%#v", valueIndex, want, got)
    99  			}
   100  		}
   101  
   102  		minValue := values[i]
   103  		maxValue := values[i]
   104  
   105  		for _, value := range values[i+1 : j] {
   106  			switch {
   107  			case typ.Compare(value, minValue) < 0:
   108  				minValue = value
   109  			case typ.Compare(value, maxValue) > 0:
   110  				maxValue = value
   111  			}
   112  		}
   113  
   114  		lowerBound, upperBound := dict.Bounds(indexes[i:j])
   115  		if !parquet.DeepEqual(lowerBound, minValue) {
   116  			t.Errorf("wrong lower bound between indexes %d and %d: want=%#v got=%#v", i, j, minValue, lowerBound)
   117  		}
   118  		if !parquet.DeepEqual(upperBound, maxValue) {
   119  			t.Errorf("wrong upper bound between indexes %d and %d: want=%#v got=%#v", i, j, maxValue, upperBound)
   120  		}
   121  
   122  		i = j
   123  	}
   124  
   125  	for i := range lookups {
   126  		lookups[i] = parquet.Value{}
   127  	}
   128  
   129  	dict.Lookup(indexes, lookups)
   130  
   131  	for lookupIndex, valueIndex := range indexes {
   132  		want := mapping[valueIndex]
   133  		got := lookups[lookupIndex]
   134  
   135  		if !parquet.Equal(want, got) {
   136  			t.Fatalf("wrong value looked up at index %d: want=%+v got=%+v", valueIndex, want, got)
   137  		}
   138  	}
   139  }
   140  
   141  func BenchmarkDictionary(b *testing.B) {
   142  	tests := []struct {
   143  		scenario string
   144  		init     func(parquet.Dictionary, []int32, []parquet.Value)
   145  		test     func(parquet.Dictionary, []int32, []parquet.Value)
   146  	}{
   147  		{
   148  			scenario: "Bounds",
   149  			init:     parquet.Dictionary.Insert,
   150  			test: func(dict parquet.Dictionary, indexes []int32, _ []parquet.Value) {
   151  				dict.Bounds(indexes)
   152  			},
   153  		},
   154  
   155  		{
   156  			scenario: "Insert",
   157  			test:     parquet.Dictionary.Insert,
   158  		},
   159  
   160  		{
   161  			scenario: "Lookup",
   162  			init:     parquet.Dictionary.Insert,
   163  			test:     parquet.Dictionary.Lookup,
   164  		},
   165  	}
   166  
   167  	for i, test := range tests {
   168  		b.Run(test.scenario, func(b *testing.B) {
   169  			for j, typ := range dictionaryTypes {
   170  				for _, numValues := range []int{1e2, 1e3, 1e4, 1e5, 1e6} {
   171  					buf := typ.NewValues(make([]byte, 0, 4*numValues), nil)
   172  					dict := typ.NewDictionary(0, 0, buf)
   173  					values := make([]parquet.Value, numValues)
   174  
   175  					f := randValueFuncOf(typ)
   176  					r := rand.New(rand.NewSource(int64(i * j * numValues)))
   177  
   178  					for i := range values {
   179  						values[i] = f(r)
   180  					}
   181  
   182  					indexes := make([]int32, len(values))
   183  					if test.init != nil {
   184  						test.init(dict, indexes, values)
   185  					}
   186  
   187  					b.Run(fmt.Sprintf("%s/N=%d", typ, numValues), func(b *testing.B) {
   188  						start := time.Now()
   189  
   190  						for i := 0; i < b.N; i++ {
   191  							test.test(dict, indexes, values)
   192  						}
   193  
   194  						seconds := time.Since(start).Seconds()
   195  						b.ReportMetric(float64(numValues*b.N)/seconds, "value/s")
   196  					})
   197  				}
   198  			}
   199  		})
   200  	}
   201  }
   202  
   203  func TestIssue312(t *testing.T) {
   204  	node := parquet.String()
   205  	node = parquet.Encoded(node, &parquet.RLEDictionary)
   206  	g := parquet.Group{}
   207  	g["mystring"] = node
   208  	schema := parquet.NewSchema("test", g)
   209  
   210  	rows := []parquet.Row{[]parquet.Value{parquet.ValueOf("hello").Level(0, 0, 0)}}
   211  
   212  	var storage bytes.Buffer
   213  
   214  	tests := []struct {
   215  		name        string
   216  		getRowGroup func(t *testing.T) parquet.RowGroup
   217  	}{
   218  		{
   219  			name: "Writer",
   220  			getRowGroup: func(t *testing.T) parquet.RowGroup {
   221  				t.Helper()
   222  
   223  				w := parquet.NewWriter(&storage, schema)
   224  				_, err := w.WriteRows(rows)
   225  				if err != nil {
   226  					t.Fatal(err)
   227  				}
   228  				if err := w.Close(); err != nil {
   229  					t.Fatal(err)
   230  				}
   231  
   232  				r := bytes.NewReader(storage.Bytes())
   233  				f, err := parquet.OpenFile(r, int64(storage.Len()))
   234  				if err != nil {
   235  					t.Fatal(err)
   236  				}
   237  				return f.RowGroups()[0]
   238  			},
   239  		},
   240  		{
   241  			name: "Buffer",
   242  			getRowGroup: func(t *testing.T) parquet.RowGroup {
   243  				t.Helper()
   244  
   245  				b := parquet.NewBuffer(schema)
   246  				_, err := b.WriteRows(rows)
   247  				if err != nil {
   248  					t.Fatal(err)
   249  				}
   250  				return b
   251  			},
   252  		},
   253  	}
   254  
   255  	for _, testCase := range tests {
   256  		t.Run(testCase.name, func(t *testing.T) {
   257  			rowGroup := testCase.getRowGroup(t)
   258  
   259  			chunk := rowGroup.ColumnChunks()[0]
   260  			idx, _ := chunk.ColumnIndex()
   261  			val := idx.MinValue(0)
   262  			columnType := chunk.Type()
   263  			values := columnType.NewValues(val.Bytes(), []uint32{0, uint32(len(val.Bytes()))})
   264  
   265  			// This test ensures that the dictionary type created by column
   266  			// chunks of parquet readers and buffers are the same. We want the
   267  			// column chunk type to be the actual value type, even when the
   268  			// schema uses a dictionary encoding.
   269  			//
   270  			// https://github.com/segmentio/parquet-go/issues/312
   271  			_ = columnType.NewDictionary(0, 1, values)
   272  		})
   273  	}
   274  }