github.com/bir3/gocompiler@v0.9.2202/src/cmd/compile/internal/ssa/nilcheck.go (about) 1 // Copyright 2015 The Go Authors. All rights reserved. 2 // Use of this source code is governed by a BSD-style 3 // license that can be found in the LICENSE file. 4 5 package ssa 6 7 import ( 8 "github.com/bir3/gocompiler/src/cmd/compile/internal/ir" 9 "github.com/bir3/gocompiler/src/cmd/internal/src" 10 "github.com/bir3/gocompiler/src/internal/buildcfg" 11 ) 12 13 // nilcheckelim eliminates unnecessary nil checks. 14 // runs on machine-independent code. 15 func nilcheckelim(f *Func) { 16 // A nil check is redundant if the same nil check was successful in a 17 // dominating block. The efficacy of this pass depends heavily on the 18 // efficacy of the cse pass. 19 sdom := f.Sdom() 20 21 // TODO: Eliminate more nil checks. 22 // We can recursively remove any chain of fixed offset calculations, 23 // i.e. struct fields and array elements, even with non-constant 24 // indices: x is non-nil iff x.a.b[i].c is. 25 26 type walkState int 27 const ( 28 Work walkState = iota // process nil checks and traverse to dominees 29 ClearPtr // forget the fact that ptr is nil 30 ) 31 32 type bp struct { 33 block *Block // block, or nil in ClearPtr state 34 ptr *Value // if non-nil, ptr that is to be cleared in ClearPtr state 35 op walkState 36 } 37 38 work := make([]bp, 0, 256) 39 work = append(work, bp{block: f.Entry}) 40 41 // map from value ID to known non-nil version of that value ID 42 // (in the current dominator path being walked). This slice is updated by 43 // walkStates to maintain the known non-nil values. 44 // If there is extrinsic information about non-nil-ness, this map 45 // points a value to itself. If a value is known non-nil because we 46 // already did a nil check on it, it points to the nil check operation. 47 nonNilValues := f.Cache.allocValueSlice(f.NumValues()) 48 defer f.Cache.freeValueSlice(nonNilValues) 49 50 // make an initial pass identifying any non-nil values 51 for _, b := range f.Blocks { 52 for _, v := range b.Values { 53 // a value resulting from taking the address of a 54 // value, or a value constructed from an offset of a 55 // non-nil ptr (OpAddPtr) implies it is non-nil 56 // We also assume unsafe pointer arithmetic generates non-nil pointers. See #27180. 57 // We assume that SlicePtr is non-nil because we do a bounds check 58 // before the slice access (and all cap>0 slices have a non-nil ptr). See #30366. 59 if v.Op == OpAddr || v.Op == OpLocalAddr || v.Op == OpAddPtr || v.Op == OpOffPtr || v.Op == OpAdd32 || v.Op == OpAdd64 || v.Op == OpSub32 || v.Op == OpSub64 || v.Op == OpSlicePtr { 60 nonNilValues[v.ID] = v 61 } 62 } 63 } 64 65 for changed := true; changed; { 66 changed = false 67 for _, b := range f.Blocks { 68 for _, v := range b.Values { 69 // phis whose arguments are all non-nil 70 // are non-nil 71 if v.Op == OpPhi { 72 argsNonNil := true 73 for _, a := range v.Args { 74 if nonNilValues[a.ID] == nil { 75 argsNonNil = false 76 break 77 } 78 } 79 if argsNonNil { 80 if nonNilValues[v.ID] == nil { 81 changed = true 82 } 83 nonNilValues[v.ID] = v 84 } 85 } 86 } 87 } 88 } 89 90 // allocate auxiliary date structures for computing store order 91 sset := f.newSparseSet(f.NumValues()) 92 defer f.retSparseSet(sset) 93 storeNumber := f.Cache.allocInt32Slice(f.NumValues()) 94 defer f.Cache.freeInt32Slice(storeNumber) 95 96 // perform a depth first walk of the dominee tree 97 for len(work) > 0 { 98 node := work[len(work)-1] 99 work = work[:len(work)-1] 100 101 switch node.op { 102 case Work: 103 b := node.block 104 105 // First, see if we're dominated by an explicit nil check. 106 if len(b.Preds) == 1 { 107 p := b.Preds[0].b 108 if p.Kind == BlockIf && p.Controls[0].Op == OpIsNonNil && p.Succs[0].b == b { 109 if ptr := p.Controls[0].Args[0]; nonNilValues[ptr.ID] == nil { 110 nonNilValues[ptr.ID] = ptr 111 work = append(work, bp{op: ClearPtr, ptr: ptr}) 112 } 113 } 114 } 115 116 // Next, order values in the current block w.r.t. stores. 117 b.Values = storeOrder(b.Values, sset, storeNumber) 118 119 pendingLines := f.cachedLineStarts // Holds statement boundaries that need to be moved to a new value/block 120 pendingLines.clear() 121 122 // Next, process values in the block. 123 for _, v := range b.Values { 124 switch v.Op { 125 case OpIsNonNil: 126 ptr := v.Args[0] 127 if nonNilValues[ptr.ID] != nil { 128 if v.Pos.IsStmt() == src.PosIsStmt { // Boolean true is a terrible statement boundary. 129 pendingLines.add(v.Pos) 130 v.Pos = v.Pos.WithNotStmt() 131 } 132 // This is a redundant explicit nil check. 133 v.reset(OpConstBool) 134 v.AuxInt = 1 // true 135 } 136 case OpNilCheck: 137 ptr := v.Args[0] 138 if nilCheck := nonNilValues[ptr.ID]; nilCheck != nil { 139 // This is a redundant implicit nil check. 140 // Logging in the style of the former compiler -- and omit line 1, 141 // which is usually in generated code. 142 if f.fe.Debug_checknil() && v.Pos.Line() > 1 { 143 f.Warnl(v.Pos, "removed nil check") 144 } 145 if v.Pos.IsStmt() == src.PosIsStmt { // About to lose a statement boundary 146 pendingLines.add(v.Pos) 147 } 148 v.Op = OpCopy 149 v.SetArgs1(nilCheck) 150 continue 151 } 152 // Record the fact that we know ptr is non nil, and remember to 153 // undo that information when this dominator subtree is done. 154 nonNilValues[ptr.ID] = v 155 work = append(work, bp{op: ClearPtr, ptr: ptr}) 156 fallthrough // a non-eliminated nil check might be a good place for a statement boundary. 157 default: 158 if v.Pos.IsStmt() != src.PosNotStmt && !isPoorStatementOp(v.Op) && pendingLines.contains(v.Pos) { 159 v.Pos = v.Pos.WithIsStmt() 160 pendingLines.remove(v.Pos) 161 } 162 } 163 } 164 // This reduces the lost statement count in "go" by 5 (out of 500 total). 165 for j := range b.Values { // is this an ordering problem? 166 v := b.Values[j] 167 if v.Pos.IsStmt() != src.PosNotStmt && !isPoorStatementOp(v.Op) && pendingLines.contains(v.Pos) { 168 v.Pos = v.Pos.WithIsStmt() 169 pendingLines.remove(v.Pos) 170 } 171 } 172 if pendingLines.contains(b.Pos) { 173 b.Pos = b.Pos.WithIsStmt() 174 pendingLines.remove(b.Pos) 175 } 176 177 // Add all dominated blocks to the work list. 178 for w := sdom[node.block.ID].child; w != nil; w = sdom[w.ID].sibling { 179 work = append(work, bp{op: Work, block: w}) 180 } 181 182 case ClearPtr: 183 nonNilValues[node.ptr.ID] = nil 184 continue 185 } 186 } 187 } 188 189 // All platforms are guaranteed to fault if we load/store to anything smaller than this address. 190 // 191 // This should agree with minLegalPointer in the runtime. 192 const minZeroPage = 4096 193 194 // faultOnLoad is true if a load to an address below minZeroPage will trigger a SIGSEGV. 195 var faultOnLoad = buildcfg.GOOS != "aix" 196 197 // nilcheckelim2 eliminates unnecessary nil checks. 198 // Runs after lowering and scheduling. 199 func nilcheckelim2(f *Func) { 200 unnecessary := f.newSparseMap(f.NumValues()) // map from pointer that will be dereferenced to index of dereferencing value in b.Values[] 201 defer f.retSparseMap(unnecessary) 202 203 pendingLines := f.cachedLineStarts // Holds statement boundaries that need to be moved to a new value/block 204 205 for _, b := range f.Blocks { 206 // Walk the block backwards. Find instructions that will fault if their 207 // input pointer is nil. Remove nil checks on those pointers, as the 208 // faulting instruction effectively does the nil check for free. 209 unnecessary.clear() 210 pendingLines.clear() 211 // Optimization: keep track of removed nilcheck with smallest index 212 firstToRemove := len(b.Values) 213 for i := len(b.Values) - 1; i >= 0; i-- { 214 v := b.Values[i] 215 if opcodeTable[v.Op].nilCheck && unnecessary.contains(v.Args[0].ID) { 216 if f.fe.Debug_checknil() && v.Pos.Line() > 1 { 217 f.Warnl(v.Pos, "removed nil check") 218 } 219 // For bug 33724, policy is that we might choose to bump an existing position 220 // off the faulting load/store in favor of the one from the nil check. 221 222 // Iteration order means that first nilcheck in the chain wins, others 223 // are bumped into the ordinary statement preservation algorithm. 224 u := b.Values[unnecessary.get(v.Args[0].ID)] 225 if !u.Pos.SameFileAndLine(v.Pos) { 226 if u.Pos.IsStmt() == src.PosIsStmt { 227 pendingLines.add(u.Pos) 228 } 229 u.Pos = v.Pos 230 } else if v.Pos.IsStmt() == src.PosIsStmt { 231 pendingLines.add(v.Pos) 232 } 233 234 v.reset(OpUnknown) 235 firstToRemove = i 236 continue 237 } 238 if v.Type.IsMemory() || v.Type.IsTuple() && v.Type.FieldType(1).IsMemory() { 239 if v.Op == OpVarLive || (v.Op == OpVarDef && !v.Aux.(*ir.Name).Type().HasPointers()) { 240 // These ops don't really change memory. 241 continue 242 // Note: OpVarDef requires that the defined variable not have pointers. 243 // We need to make sure that there's no possible faulting 244 // instruction between a VarDef and that variable being 245 // fully initialized. If there was, then anything scanning 246 // the stack during the handling of that fault will see 247 // a live but uninitialized pointer variable on the stack. 248 // 249 // If we have: 250 // 251 // NilCheck p 252 // VarDef x 253 // x = *p 254 // 255 // We can't rewrite that to 256 // 257 // VarDef x 258 // NilCheck p 259 // x = *p 260 // 261 // Particularly, even though *p faults on p==nil, we still 262 // have to do the explicit nil check before the VarDef. 263 // See issue #32288. 264 } 265 // This op changes memory. Any faulting instruction after v that 266 // we've recorded in the unnecessary map is now obsolete. 267 unnecessary.clear() 268 } 269 270 // Find any pointers that this op is guaranteed to fault on if nil. 271 var ptrstore [2]*Value 272 ptrs := ptrstore[:0] 273 if opcodeTable[v.Op].faultOnNilArg0 && (faultOnLoad || v.Type.IsMemory()) { 274 // On AIX, only writing will fault. 275 ptrs = append(ptrs, v.Args[0]) 276 } 277 if opcodeTable[v.Op].faultOnNilArg1 && (faultOnLoad || (v.Type.IsMemory() && v.Op != OpPPC64LoweredMove)) { 278 // On AIX, only writing will fault. 279 // LoweredMove is a special case because it's considered as a "mem" as it stores on arg0 but arg1 is accessed as a load and should be checked. 280 ptrs = append(ptrs, v.Args[1]) 281 } 282 283 for _, ptr := range ptrs { 284 // Check to make sure the offset is small. 285 switch opcodeTable[v.Op].auxType { 286 case auxSym: 287 if v.Aux != nil { 288 continue 289 } 290 case auxSymOff: 291 if v.Aux != nil || v.AuxInt < 0 || v.AuxInt >= minZeroPage { 292 continue 293 } 294 case auxSymValAndOff: 295 off := ValAndOff(v.AuxInt).Off() 296 if v.Aux != nil || off < 0 || off >= minZeroPage { 297 continue 298 } 299 case auxInt32: 300 // Mips uses this auxType for atomic add constant. It does not affect the effective address. 301 case auxInt64: 302 // ARM uses this auxType for duffcopy/duffzero/alignment info. 303 // It does not affect the effective address. 304 case auxNone: 305 // offset is zero. 306 default: 307 v.Fatalf("can't handle aux %s (type %d) yet\n", v.auxString(), int(opcodeTable[v.Op].auxType)) 308 } 309 // This instruction is guaranteed to fault if ptr is nil. 310 // Any previous nil check op is unnecessary. 311 unnecessary.set(ptr.ID, int32(i)) 312 } 313 } 314 // Remove values we've clobbered with OpUnknown. 315 i := firstToRemove 316 for j := i; j < len(b.Values); j++ { 317 v := b.Values[j] 318 if v.Op != OpUnknown { 319 if !notStmtBoundary(v.Op) && pendingLines.contains(v.Pos) { // Late in compilation, so any remaining NotStmt values are probably okay now. 320 v.Pos = v.Pos.WithIsStmt() 321 pendingLines.remove(v.Pos) 322 } 323 b.Values[i] = v 324 i++ 325 } 326 } 327 328 if pendingLines.contains(b.Pos) { 329 b.Pos = b.Pos.WithIsStmt() 330 } 331 332 b.truncateValues(i) 333 334 // TODO: if b.Kind == BlockPlain, start the analysis in the subsequent block to find 335 // more unnecessary nil checks. Would fix test/nilptr3.go:159. 336 } 337 }