github.com/apache/arrow/go/v14@v14.0.1/parquet/pqarrow/path_builder.go (about)

     1  // Licensed to the Apache Software Foundation (ASF) under one
     2  // or more contributor license agreements.  See the NOTICE file
     3  // distributed with this work for additional information
     4  // regarding copyright ownership.  The ASF licenses this file
     5  // to you under the Apache License, Version 2.0 (the
     6  // "License"); you may not use this file except in compliance
     7  // with the License.  You may obtain a copy of the License at
     8  //
     9  // http://www.apache.org/licenses/LICENSE-2.0
    10  //
    11  // Unless required by applicable law or agreed to in writing, software
    12  // distributed under the License is distributed on an "AS IS" BASIS,
    13  // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
    14  // See the License for the specific language governing permissions and
    15  // limitations under the License.
    16  
    17  package pqarrow
    18  
    19  import (
    20  	"fmt"
    21  	"sync/atomic"
    22  	"unsafe"
    23  
    24  	"github.com/apache/arrow/go/v14/arrow"
    25  	"github.com/apache/arrow/go/v14/arrow/array"
    26  	"github.com/apache/arrow/go/v14/arrow/memory"
    27  	"github.com/apache/arrow/go/v14/internal/bitutils"
    28  	"github.com/apache/arrow/go/v14/parquet/internal/encoding"
    29  	"golang.org/x/xerrors"
    30  )
    31  
    32  type iterResult int8
    33  
    34  const (
    35  	iterDone iterResult = -1
    36  	iterNext iterResult = 1
    37  )
    38  
    39  type elemRange struct {
    40  	start int64
    41  	end   int64
    42  }
    43  
    44  func (e elemRange) empty() bool { return e.start == e.end }
    45  func (e elemRange) size() int64 { return e.end - e.start }
    46  
    47  type rangeSelector interface {
    48  	GetRange(idx int64) elemRange
    49  }
    50  
    51  type varRangeSelector struct {
    52  	offsets []int32
    53  }
    54  
    55  func (v varRangeSelector) GetRange(idx int64) elemRange {
    56  	return elemRange{int64(v.offsets[idx]), int64(v.offsets[idx+1])}
    57  }
    58  
    59  type fixedSizeRangeSelector struct {
    60  	listSize int32
    61  }
    62  
    63  func (f fixedSizeRangeSelector) GetRange(idx int64) elemRange {
    64  	start := idx * int64(f.listSize)
    65  	return elemRange{start, start + int64(f.listSize)}
    66  }
    67  
    68  type pathNode interface {
    69  	clone() pathNode
    70  }
    71  
    72  type allPresentTerminalNode struct {
    73  	defLevel int16
    74  }
    75  
    76  func (n *allPresentTerminalNode) clone() pathNode {
    77  	ret := *n
    78  	return &ret
    79  }
    80  
    81  func (n *allPresentTerminalNode) run(rng elemRange, ctx *pathWriteCtx) iterResult {
    82  	return ctx.AppendDefLevels(int(rng.size()), n.defLevel)
    83  }
    84  
    85  type allNullsTerminalNode struct {
    86  	defLevel int16
    87  	repLevel int16
    88  }
    89  
    90  func (n *allNullsTerminalNode) clone() pathNode {
    91  	ret := *n
    92  	return &ret
    93  }
    94  
    95  func (n *allNullsTerminalNode) run(rng elemRange, ctx *pathWriteCtx) iterResult {
    96  	fillRepLevels(int(rng.size()), n.repLevel, ctx)
    97  	return ctx.AppendDefLevels(int(rng.size()), n.defLevel)
    98  }
    99  
   100  type nullableTerminalNode struct {
   101  	bitmap            []byte
   102  	elemOffset        int64
   103  	defLevelIfPresent int16
   104  	defLevelIfNull    int16
   105  }
   106  
   107  func (n *nullableTerminalNode) clone() pathNode {
   108  	ret := *n
   109  	return &ret
   110  }
   111  
   112  func (n *nullableTerminalNode) run(rng elemRange, ctx *pathWriteCtx) iterResult {
   113  	elems := rng.size()
   114  	ctx.ReserveDefLevels(int(elems))
   115  
   116  	var (
   117  		present = (*(*[2]byte)(unsafe.Pointer(&n.defLevelIfPresent)))[:]
   118  		null    = (*(*[2]byte)(unsafe.Pointer(&n.defLevelIfNull)))[:]
   119  	)
   120  	rdr := bitutils.NewBitRunReader(n.bitmap, n.elemOffset+rng.start, elems)
   121  	for {
   122  		run := rdr.NextRun()
   123  		if run.Len == 0 {
   124  			break
   125  		}
   126  		if run.Set {
   127  			ctx.defLevels.UnsafeWriteCopy(int(run.Len), present)
   128  		} else {
   129  			ctx.defLevels.UnsafeWriteCopy(int(run.Len), null)
   130  		}
   131  	}
   132  	return iterDone
   133  }
   134  
   135  type listNode struct {
   136  	selector        rangeSelector
   137  	prevRepLevel    int16
   138  	repLevel        int16
   139  	defLevelIfEmpty int16
   140  	isLast          bool
   141  }
   142  
   143  func (n *listNode) clone() pathNode {
   144  	ret := *n
   145  	return &ret
   146  }
   147  
   148  func (n *listNode) run(rng, childRng *elemRange, ctx *pathWriteCtx) iterResult {
   149  	if rng.empty() {
   150  		return iterDone
   151  	}
   152  
   153  	// find the first non-empty list (skipping a run of empties)
   154  	start := rng.start
   155  	for {
   156  		// retrieve the range of elements that this list contains
   157  		*childRng = n.selector.GetRange(rng.start)
   158  		if !childRng.empty() {
   159  			break
   160  		}
   161  		rng.start++
   162  		if rng.empty() {
   163  			break
   164  		}
   165  	}
   166  
   167  	// loops post-condition:
   168  	// * rng is either empty (we're done processing this node)
   169  	//     or start corresponds to a non-empty list
   170  	// * if rng is non-empty, childRng contains the bounds of the non-empty list
   171  
   172  	// handle any skipped over empty lists
   173  	emptyElems := rng.start - start
   174  	if emptyElems > 0 {
   175  		fillRepLevels(int(emptyElems), n.prevRepLevel, ctx)
   176  		ctx.AppendDefLevels(int(emptyElems), n.defLevelIfEmpty)
   177  	}
   178  
   179  	// start of a new list, note that for nested lists adding the element
   180  	// here effectively suppresses this code until we either encounter null
   181  	// elements or empty lists between here and the innermost list (since we
   182  	// make the rep levels repetition and definition levels unequal).
   183  	// similarly when we are backtracking up the stack, the repetition
   184  	// and definition levels are again equal so if we encounter an intermediate
   185  	// list, with more elements, this will detect it as a new list
   186  	if ctx.equalRepDeflevlsLen() && !rng.empty() {
   187  		ctx.AppendRepLevel(n.prevRepLevel)
   188  	}
   189  
   190  	if rng.empty() {
   191  		return iterDone
   192  	}
   193  
   194  	rng.start++
   195  	if n.isLast {
   196  		// if this is the last repeated node, we can try
   197  		// to extend the child range as wide as possible,
   198  		// before continuing to the next node
   199  		return n.fillForLast(rng, childRng, ctx)
   200  	}
   201  
   202  	return iterNext
   203  }
   204  
   205  func (n *listNode) fillForLast(rng, childRng *elemRange, ctx *pathWriteCtx) iterResult {
   206  	fillRepLevels(int(childRng.size()), n.repLevel, ctx)
   207  	// once we've reached this point the following preconditions should hold:
   208  	// 1. there are no more repeated path nodes to deal with
   209  	// 2. all elements in |range| reperesent contiguous elements in the child
   210  	//    array (null values would have shortened the range to ensure all
   211  	//    remaining list elements are present, though they may be empty)
   212  	// 3. no element of range spans a parent list (intermediate list nodes
   213  	//    only handle one list entry at a time)
   214  	//
   215  	// given these preconditions, it should be safe to fill runs on non-empty lists
   216  	// here and expand the range in the child node accordingly
   217  	for !rng.empty() {
   218  		sizeCheck := n.selector.GetRange(rng.start)
   219  		if sizeCheck.empty() {
   220  			// the empty range will need to be handled after we pass down the accumulated
   221  			// range because it affects def level placement and we need to get the children
   222  			// def levels entered first
   223  			break
   224  		}
   225  
   226  		// this is the start of a new list. we can be sure that it only applies to the
   227  		// previous list (and doesn't jump to the start of any list further up in nesting
   228  		// due to the contraints mentioned earlier)
   229  		ctx.AppendRepLevel(n.prevRepLevel)
   230  		ctx.AppendRepLevels(int(sizeCheck.size())-1, n.repLevel)
   231  		childRng.end = sizeCheck.end
   232  		rng.start++
   233  	}
   234  
   235  	// do book-keeping to track the elements of the arrays that are actually visited
   236  	// beyond this point. this is necessary to identify "gaps" in values that should
   237  	// not be processed (written out to parquet)
   238  	ctx.recordPostListVisit(*childRng)
   239  	return iterNext
   240  }
   241  
   242  type nullableNode struct {
   243  	bitmap         []byte
   244  	entryOffset    int64
   245  	repLevelIfNull int16
   246  	defLevelIfNull int16
   247  
   248  	validBitsReader bitutils.BitRunReader
   249  	newRange        bool
   250  }
   251  
   252  func (n *nullableNode) clone() pathNode {
   253  	var ret nullableNode = *n
   254  	return &ret
   255  }
   256  
   257  func (n *nullableNode) run(rng, childRng *elemRange, ctx *pathWriteCtx) iterResult {
   258  	if n.newRange {
   259  		n.validBitsReader = bitutils.NewBitRunReader(n.bitmap, n.entryOffset+rng.start, rng.size())
   260  	}
   261  	childRng.start = rng.start
   262  	run := n.validBitsReader.NextRun()
   263  	if !run.Set {
   264  		rng.start += run.Len
   265  		fillRepLevels(int(run.Len), n.repLevelIfNull, ctx)
   266  		ctx.AppendDefLevels(int(run.Len), n.defLevelIfNull)
   267  		run = n.validBitsReader.NextRun()
   268  	}
   269  
   270  	if rng.empty() {
   271  		n.newRange = true
   272  		return iterDone
   273  	}
   274  	childRng.start = rng.start
   275  	childRng.end = childRng.start
   276  	childRng.end += run.Len
   277  	rng.start += childRng.size()
   278  	n.newRange = false
   279  	return iterNext
   280  }
   281  
   282  type pathInfo struct {
   283  	path           []pathNode
   284  	primitiveArr   arrow.Array
   285  	maxDefLevel    int16
   286  	maxRepLevel    int16
   287  	leafIsNullable bool
   288  }
   289  
   290  func (p pathInfo) clone() pathInfo {
   291  	ret := p
   292  	ret.path = make([]pathNode, len(p.path))
   293  	for idx, n := range p.path {
   294  		ret.path[idx] = n.clone()
   295  	}
   296  	return ret
   297  }
   298  
   299  type pathBuilder struct {
   300  	info             pathInfo
   301  	paths            []pathInfo
   302  	nullableInParent bool
   303  
   304  	refCount int64
   305  }
   306  
   307  func (p *pathBuilder) Retain() {
   308  	atomic.AddInt64(&p.refCount, 1)
   309  }
   310  
   311  func (p *pathBuilder) Release() {
   312  	if atomic.AddInt64(&p.refCount, -1) == 0 {
   313  		for idx := range p.paths {
   314  			p.paths[idx].primitiveArr.Release()
   315  			p.paths[idx].primitiveArr = nil
   316  		}
   317  	}
   318  }
   319  
   320  // calling NullN on the arr directly will compute the nulls
   321  // if we have "UnknownNullCount", calling NullN on the data
   322  // object directly will just return the value the data has.
   323  // thus we might bet array.UnknownNullCount as the result here.
   324  func lazyNullCount(arr arrow.Array) int64 {
   325  	return int64(arr.Data().NullN())
   326  }
   327  
   328  func lazyNoNulls(arr arrow.Array) bool {
   329  	nulls := lazyNullCount(arr)
   330  	return nulls == 0 || (nulls == array.UnknownNullCount && arr.NullBitmapBytes() == nil)
   331  }
   332  
   333  type fixupVisitor struct {
   334  	maxRepLevel    int
   335  	repLevelIfNull int16
   336  }
   337  
   338  func (f *fixupVisitor) visit(n pathNode) {
   339  	switch n := n.(type) {
   340  	case *listNode:
   341  		if n.repLevel == int16(f.maxRepLevel) {
   342  			n.isLast = true
   343  			f.repLevelIfNull = -1
   344  		} else {
   345  			f.repLevelIfNull = n.repLevel
   346  		}
   347  	case *nullableTerminalNode:
   348  	case *allPresentTerminalNode:
   349  	case *allNullsTerminalNode:
   350  		if f.repLevelIfNull != -1 {
   351  			n.repLevel = f.repLevelIfNull
   352  		}
   353  	case *nullableNode:
   354  		if f.repLevelIfNull != -1 {
   355  			n.repLevelIfNull = f.repLevelIfNull
   356  		}
   357  	}
   358  }
   359  
   360  func fixup(info pathInfo) pathInfo {
   361  	// we only need to fixup the path if there were repeated elems
   362  	if info.maxRepLevel == 0 {
   363  		return info
   364  	}
   365  
   366  	visitor := fixupVisitor{maxRepLevel: int(info.maxRepLevel)}
   367  	if visitor.maxRepLevel > 0 {
   368  		visitor.repLevelIfNull = 0
   369  	} else {
   370  		visitor.repLevelIfNull = -1
   371  	}
   372  
   373  	for _, p := range info.path {
   374  		visitor.visit(p)
   375  	}
   376  	return info
   377  }
   378  
   379  func (p *pathBuilder) Visit(arr arrow.Array) error {
   380  	switch arr.DataType().ID() {
   381  	case arrow.LIST, arrow.MAP:
   382  		p.maybeAddNullable(arr)
   383  		// increment necessary due to empty lists
   384  		p.info.maxDefLevel++
   385  		p.info.maxRepLevel++
   386  		larr, ok := arr.(*array.List)
   387  		if !ok {
   388  			larr = arr.(*array.Map).List
   389  		}
   390  
   391  		p.info.path = append(p.info.path, &listNode{
   392  			selector:        varRangeSelector{larr.Offsets()[larr.Data().Offset():]},
   393  			prevRepLevel:    p.info.maxRepLevel - 1,
   394  			repLevel:        p.info.maxRepLevel,
   395  			defLevelIfEmpty: p.info.maxDefLevel - 1,
   396  		})
   397  		p.nullableInParent = ok
   398  		return p.Visit(larr.ListValues())
   399  	case arrow.FIXED_SIZE_LIST:
   400  		p.maybeAddNullable(arr)
   401  		larr := arr.(*array.FixedSizeList)
   402  		listSize := larr.DataType().(*arrow.FixedSizeListType).Len()
   403  		// technically we could encode fixed sized lists with two level encodings
   404  		// but we always use 3 level encoding, so we increment def levels as well
   405  		p.info.maxDefLevel++
   406  		p.info.maxRepLevel++
   407  		p.info.path = append(p.info.path, &listNode{
   408  			selector:        fixedSizeRangeSelector{listSize},
   409  			prevRepLevel:    p.info.maxRepLevel - 1,
   410  			repLevel:        p.info.maxRepLevel,
   411  			defLevelIfEmpty: p.info.maxDefLevel,
   412  		})
   413  		// if arr.data.offset > 0, slice?
   414  		return p.Visit(larr.ListValues())
   415  	case arrow.DICTIONARY:
   416  		// only currently handle dictionaryarray where the dictionary
   417  		// is a primitive type
   418  		dictArr := arr.(*array.Dictionary)
   419  		valType := dictArr.DataType().(*arrow.DictionaryType).ValueType
   420  		if _, ok := valType.(arrow.NestedType); ok {
   421  			return fmt.Errorf("%w: writing DictionaryArray with nested dictionary type not yet supported",
   422  				arrow.ErrNotImplemented)
   423  		}
   424  		if dictArr.Dictionary().NullN() > 0 {
   425  			return fmt.Errorf("%w: writing DictionaryArray with null encoded in dictionary not yet supported",
   426  				arrow.ErrNotImplemented)
   427  		}
   428  		p.addTerminalInfo(arr)
   429  		return nil
   430  	case arrow.STRUCT:
   431  		p.maybeAddNullable(arr)
   432  		infoBackup := p.info
   433  		dt := arr.DataType().(*arrow.StructType)
   434  		for idx, f := range dt.Fields() {
   435  			p.nullableInParent = f.Nullable
   436  			if err := p.Visit(arr.(*array.Struct).Field(idx)); err != nil {
   437  				return err
   438  			}
   439  			p.info = infoBackup
   440  		}
   441  		return nil
   442  	case arrow.EXTENSION:
   443  		return p.Visit(arr.(array.ExtensionArray).Storage())
   444  	case arrow.SPARSE_UNION, arrow.DENSE_UNION:
   445  		return xerrors.New("union types aren't supported in parquet")
   446  	default:
   447  		p.addTerminalInfo(arr)
   448  		return nil
   449  	}
   450  }
   451  
   452  func (p *pathBuilder) addTerminalInfo(arr arrow.Array) {
   453  	p.info.leafIsNullable = p.nullableInParent
   454  	if p.nullableInParent {
   455  		p.info.maxDefLevel++
   456  	}
   457  
   458  	// we don't use null_count because if the null_count isn't known
   459  	// and the array does in fact contain nulls, we will end up traversing
   460  	// the null bitmap twice.
   461  	if lazyNoNulls(arr) {
   462  		p.info.path = append(p.info.path, &allPresentTerminalNode{p.info.maxDefLevel})
   463  		p.info.leafIsNullable = false
   464  	} else if lazyNullCount(arr) == int64(arr.Len()) {
   465  		p.info.path = append(p.info.path, &allNullsTerminalNode{p.info.maxDefLevel - 1, -1})
   466  	} else {
   467  		p.info.path = append(p.info.path, &nullableTerminalNode{bitmap: arr.NullBitmapBytes(), elemOffset: int64(arr.Data().Offset()), defLevelIfPresent: p.info.maxDefLevel, defLevelIfNull: p.info.maxDefLevel - 1})
   468  	}
   469  	arr.Retain()
   470  	p.info.primitiveArr = arr
   471  	p.paths = append(p.paths, fixup(p.info.clone()))
   472  }
   473  
   474  func (p *pathBuilder) maybeAddNullable(arr arrow.Array) {
   475  	if !p.nullableInParent {
   476  		return
   477  	}
   478  
   479  	p.info.maxDefLevel++
   480  	if lazyNoNulls(arr) {
   481  		return
   482  	}
   483  
   484  	if lazyNullCount(arr) == int64(arr.Len()) {
   485  		p.info.path = append(p.info.path, &allNullsTerminalNode{p.info.maxDefLevel - 1, -1})
   486  		return
   487  	}
   488  
   489  	p.info.path = append(p.info.path, &nullableNode{
   490  		bitmap: arr.NullBitmapBytes(), entryOffset: int64(arr.Data().Offset()),
   491  		defLevelIfNull: p.info.maxDefLevel - 1, repLevelIfNull: -1,
   492  		newRange: true,
   493  	})
   494  }
   495  
   496  type multipathLevelBuilder struct {
   497  	rootRange elemRange
   498  	data      arrow.ArrayData
   499  	builder   pathBuilder
   500  
   501  	refCount int64
   502  }
   503  
   504  func (m *multipathLevelBuilder) Retain() {
   505  	atomic.AddInt64(&m.refCount, 1)
   506  }
   507  
   508  func (m *multipathLevelBuilder) Release() {
   509  	if atomic.AddInt64(&m.refCount, -1) == 0 {
   510  		m.data.Release()
   511  		m.data = nil
   512  		m.builder.Release()
   513  		m.builder = pathBuilder{}
   514  	}
   515  }
   516  
   517  func newMultipathLevelBuilder(arr arrow.Array, fieldNullable bool) (*multipathLevelBuilder, error) {
   518  	ret := &multipathLevelBuilder{
   519  		refCount:  1,
   520  		rootRange: elemRange{int64(0), int64(arr.Data().Len())},
   521  		data:      arr.Data(),
   522  		builder:   pathBuilder{nullableInParent: fieldNullable, paths: make([]pathInfo, 0), refCount: 1},
   523  	}
   524  	if err := ret.builder.Visit(arr); err != nil {
   525  		return nil, err
   526  	}
   527  	arr.Data().Retain()
   528  	return ret, nil
   529  }
   530  
   531  func (m *multipathLevelBuilder) leafCount() int {
   532  	return len(m.builder.paths)
   533  }
   534  
   535  func (m *multipathLevelBuilder) write(leafIdx int, ctx *arrowWriteContext) (multipathLevelResult, error) {
   536  	return writePath(m.rootRange, &m.builder.paths[leafIdx], ctx)
   537  }
   538  
   539  func (m *multipathLevelBuilder) writeAll(ctx *arrowWriteContext) (res []multipathLevelResult, err error) {
   540  	res = make([]multipathLevelResult, m.leafCount())
   541  	for idx := range res {
   542  		res[idx], err = m.write(idx, ctx)
   543  		if err != nil {
   544  			break
   545  		}
   546  	}
   547  	return
   548  }
   549  
   550  type multipathLevelResult struct {
   551  	leafArr         arrow.Array
   552  	defLevels       []int16
   553  	defLevelsBuffer encoding.Buffer
   554  	repLevels       []int16
   555  	repLevelsBuffer encoding.Buffer
   556  	// contains the element ranges of the required visiting on the descendants of the
   557  	// final list ancestor for any leaf node.
   558  	//
   559  	// the algorithm will attempt to consolidate the visited ranges into the smallest number
   560  	//
   561  	// this data is necessary to pass along because after producing the def-rep levels for each
   562  	// leaf array, it is impossible to determine which values have to be sent to parquet when a
   563  	// null list value in a nullable listarray is non-empty
   564  	//
   565  	// this allows for the parquet writing to determine which values ultimately need to be written
   566  	postListVisitedElems []elemRange
   567  
   568  	leafIsNullable bool
   569  }
   570  
   571  func (m *multipathLevelResult) Release() {
   572  	m.defLevels = nil
   573  	if m.defLevelsBuffer != nil {
   574  		m.defLevelsBuffer.Release()
   575  	}
   576  	if m.repLevels != nil {
   577  		m.repLevels = nil
   578  		m.repLevelsBuffer.Release()
   579  	}
   580  }
   581  
   582  type pathWriteCtx struct {
   583  	mem          memory.Allocator
   584  	defLevels    *int16BufferBuilder
   585  	repLevels    *int16BufferBuilder
   586  	visitedElems []elemRange
   587  }
   588  
   589  func (p *pathWriteCtx) ReserveDefLevels(elems int) iterResult {
   590  	p.defLevels.Reserve(elems)
   591  	return iterDone
   592  }
   593  
   594  func (p *pathWriteCtx) AppendDefLevel(lvl int16) iterResult {
   595  	p.defLevels.Append(lvl)
   596  	return iterDone
   597  }
   598  
   599  func (p *pathWriteCtx) AppendDefLevels(count int, defLevel int16) iterResult {
   600  	p.defLevels.AppendCopies(count, defLevel)
   601  	return iterDone
   602  }
   603  
   604  func (p *pathWriteCtx) UnsafeAppendDefLevel(v int16) iterResult {
   605  	p.defLevels.UnsafeAppend(v)
   606  	return iterDone
   607  }
   608  
   609  func (p *pathWriteCtx) AppendRepLevel(lvl int16) iterResult {
   610  	p.repLevels.Append(lvl)
   611  	return iterDone
   612  }
   613  
   614  func (p *pathWriteCtx) AppendRepLevels(count int, lvl int16) iterResult {
   615  	p.repLevels.AppendCopies(count, lvl)
   616  	return iterDone
   617  }
   618  
   619  func (p *pathWriteCtx) equalRepDeflevlsLen() bool { return p.defLevels.Len() == p.repLevels.Len() }
   620  
   621  func (p *pathWriteCtx) recordPostListVisit(rng elemRange) {
   622  	if len(p.visitedElems) > 0 && rng.start == p.visitedElems[len(p.visitedElems)-1].end {
   623  		p.visitedElems[len(p.visitedElems)-1].end = rng.end
   624  		return
   625  	}
   626  	p.visitedElems = append(p.visitedElems, rng)
   627  }
   628  
   629  type int16BufferBuilder struct {
   630  	*encoding.PooledBufferWriter
   631  }
   632  
   633  func (b *int16BufferBuilder) Values() []int16 {
   634  	return arrow.Int16Traits.CastFromBytes(b.PooledBufferWriter.Bytes())
   635  }
   636  
   637  func (b *int16BufferBuilder) Value(i int) int16 {
   638  	return b.Values()[i]
   639  }
   640  
   641  func (b *int16BufferBuilder) Reserve(n int) {
   642  	b.PooledBufferWriter.Reserve(n * arrow.Int16SizeBytes)
   643  }
   644  
   645  func (b *int16BufferBuilder) Len() int { return b.PooledBufferWriter.Len() / arrow.Int16SizeBytes }
   646  
   647  func (b *int16BufferBuilder) AppendCopies(count int, val int16) {
   648  	b.Reserve(count)
   649  	b.UnsafeWriteCopy(count, (*(*[2]byte)(unsafe.Pointer(&val)))[:])
   650  }
   651  
   652  func (b *int16BufferBuilder) UnsafeAppend(v int16) {
   653  	b.PooledBufferWriter.UnsafeWrite((*(*[2]byte)(unsafe.Pointer(&v)))[:])
   654  }
   655  
   656  func (b *int16BufferBuilder) Append(v int16) {
   657  	b.PooledBufferWriter.Reserve(arrow.Int16SizeBytes)
   658  	b.PooledBufferWriter.Write((*(*[2]byte)(unsafe.Pointer(&v)))[:])
   659  }
   660  
   661  func fillRepLevels(count int, repLvl int16, ctx *pathWriteCtx) {
   662  	if repLvl == -1 {
   663  		return
   664  	}
   665  
   666  	fillCount := count
   667  	// this condition occurs (rep and def levels equals), in one of a few cases:
   668  	// 1. before any list is encounted
   669  	// 2. after rep-level has been filled in due to null/empty values above
   670  	// 3. after finishing a list
   671  	if !ctx.equalRepDeflevlsLen() {
   672  		fillCount--
   673  	}
   674  	ctx.AppendRepLevels(fillCount, repLvl)
   675  }
   676  
   677  func writePath(rootRange elemRange, info *pathInfo, arrCtx *arrowWriteContext) (multipathLevelResult, error) {
   678  	stack := make([]elemRange, len(info.path))
   679  	buildResult := multipathLevelResult{
   680  		leafArr:        info.primitiveArr,
   681  		leafIsNullable: info.leafIsNullable,
   682  	}
   683  
   684  	if info.maxDefLevel == 0 {
   685  		// this case only occurs when there are no nullable or repeated columns in the path from the root to the leaf
   686  		leafLen := buildResult.leafArr.Len()
   687  		buildResult.postListVisitedElems = []elemRange{{0, int64(leafLen)}}
   688  		return buildResult, nil
   689  	}
   690  
   691  	stack[0] = rootRange
   692  	if arrCtx.defLevelsBuffer != nil {
   693  		arrCtx.defLevelsBuffer.Release()
   694  		arrCtx.defLevelsBuffer = nil
   695  	}
   696  	if arrCtx.repLevelsBuffer != nil {
   697  		arrCtx.repLevelsBuffer.Release()
   698  		arrCtx.repLevelsBuffer = nil
   699  	}
   700  
   701  	ctx := pathWriteCtx{arrCtx.props.mem,
   702  		&int16BufferBuilder{encoding.NewPooledBufferWriter(0)},
   703  		&int16BufferBuilder{encoding.NewPooledBufferWriter(0)},
   704  		make([]elemRange, 0)}
   705  
   706  	ctx.defLevels.Reserve(int(rootRange.size()))
   707  	if info.maxRepLevel > 0 {
   708  		ctx.repLevels.Reserve(int(rootRange.size()))
   709  	}
   710  
   711  	stackBase := 0
   712  	stackPos := stackBase
   713  	for stackPos >= stackBase {
   714  		var res iterResult
   715  		switch n := info.path[stackPos].(type) {
   716  		case *nullableNode:
   717  			res = n.run(&stack[stackPos], &stack[stackPos+1], &ctx)
   718  		case *listNode:
   719  			res = n.run(&stack[stackPos], &stack[stackPos+1], &ctx)
   720  		case *nullableTerminalNode:
   721  			res = n.run(stack[stackPos], &ctx)
   722  		case *allPresentTerminalNode:
   723  			res = n.run(stack[stackPos], &ctx)
   724  		case *allNullsTerminalNode:
   725  			res = n.run(stack[stackPos], &ctx)
   726  		}
   727  		stackPos += int(res)
   728  	}
   729  
   730  	if ctx.repLevels.Len() > 0 {
   731  		// this case only occurs when there was a repeated element somewhere
   732  		buildResult.repLevels = ctx.repLevels.Values()
   733  		buildResult.repLevelsBuffer = ctx.repLevels.Finish()
   734  
   735  		buildResult.postListVisitedElems, ctx.visitedElems = ctx.visitedElems, buildResult.postListVisitedElems
   736  		// if it is possible when processing lists that all lists were empty. in this
   737  		// case, no elements would have been added to the postListVisitedElements. by
   738  		// adding an empty element, we avoid special casing later
   739  		if len(buildResult.postListVisitedElems) == 0 {
   740  			buildResult.postListVisitedElems = append(buildResult.postListVisitedElems, elemRange{0, 0})
   741  		}
   742  	} else {
   743  		buildResult.postListVisitedElems = append(buildResult.postListVisitedElems, elemRange{0, int64(buildResult.leafArr.Len())})
   744  		buildResult.repLevels = nil
   745  	}
   746  
   747  	buildResult.defLevels = ctx.defLevels.Values()
   748  	buildResult.defLevelsBuffer = ctx.defLevels.Finish()
   749  	return buildResult, nil
   750  }