github.com/c0deoo1/golang1.5@v0.0.0-20220525150107-c87c805d4593/src/cmd/compile/internal/ppc64/ggen.go (about) 1 // Copyright 2009 The Go Authors. All rights reserved. 2 // Use of this source code is governed by a BSD-style 3 // license that can be found in the LICENSE file. 4 5 package ppc64 6 7 import ( 8 "cmd/compile/internal/gc" 9 "cmd/internal/obj" 10 "cmd/internal/obj/ppc64" 11 "fmt" 12 ) 13 14 func defframe(ptxt *obj.Prog) { 15 var n *gc.Node 16 17 // fill in argument size, stack size 18 ptxt.To.Type = obj.TYPE_TEXTSIZE 19 20 ptxt.To.Val = int32(gc.Rnd(gc.Curfn.Type.Argwid, int64(gc.Widthptr))) 21 frame := uint32(gc.Rnd(gc.Stksize+gc.Maxarg, int64(gc.Widthreg))) 22 ptxt.To.Offset = int64(frame) 23 24 // insert code to zero ambiguously live variables 25 // so that the garbage collector only sees initialized values 26 // when it looks for pointers. 27 p := ptxt 28 29 hi := int64(0) 30 lo := hi 31 32 // iterate through declarations - they are sorted in decreasing xoffset order. 33 for l := gc.Curfn.Func.Dcl; l != nil; l = l.Next { 34 n = l.N 35 if !n.Name.Needzero { 36 continue 37 } 38 if n.Class != gc.PAUTO { 39 gc.Fatal("needzero class %d", n.Class) 40 } 41 if n.Type.Width%int64(gc.Widthptr) != 0 || n.Xoffset%int64(gc.Widthptr) != 0 || n.Type.Width == 0 { 42 gc.Fatal("var %v has size %d offset %d", gc.Nconv(n, obj.FmtLong), int(n.Type.Width), int(n.Xoffset)) 43 } 44 45 if lo != hi && n.Xoffset+n.Type.Width >= lo-int64(2*gc.Widthreg) { 46 // merge with range we already have 47 lo = n.Xoffset 48 49 continue 50 } 51 52 // zero old range 53 p = zerorange(p, int64(frame), lo, hi) 54 55 // set new range 56 hi = n.Xoffset + n.Type.Width 57 58 lo = n.Xoffset 59 } 60 61 // zero final range 62 zerorange(p, int64(frame), lo, hi) 63 } 64 65 func zerorange(p *obj.Prog, frame int64, lo int64, hi int64) *obj.Prog { 66 cnt := hi - lo 67 if cnt == 0 { 68 return p 69 } 70 if cnt < int64(4*gc.Widthptr) { 71 for i := int64(0); i < cnt; i += int64(gc.Widthptr) { 72 p = appendpp(p, ppc64.AMOVD, obj.TYPE_REG, ppc64.REGZERO, 0, obj.TYPE_MEM, ppc64.REGSP, 8+frame+lo+i) 73 } 74 } else if cnt <= int64(128*gc.Widthptr) { 75 p = appendpp(p, ppc64.AADD, obj.TYPE_CONST, 0, 8+frame+lo-8, obj.TYPE_REG, ppc64.REGRT1, 0) 76 p.Reg = ppc64.REGSP 77 p = appendpp(p, obj.ADUFFZERO, obj.TYPE_NONE, 0, 0, obj.TYPE_MEM, 0, 0) 78 f := gc.Sysfunc("duffzero") 79 gc.Naddr(&p.To, f) 80 gc.Afunclit(&p.To, f) 81 p.To.Offset = 4 * (128 - cnt/int64(gc.Widthptr)) 82 } else { 83 p = appendpp(p, ppc64.AMOVD, obj.TYPE_CONST, 0, 8+frame+lo-8, obj.TYPE_REG, ppc64.REGTMP, 0) 84 p = appendpp(p, ppc64.AADD, obj.TYPE_REG, ppc64.REGTMP, 0, obj.TYPE_REG, ppc64.REGRT1, 0) 85 p.Reg = ppc64.REGSP 86 p = appendpp(p, ppc64.AMOVD, obj.TYPE_CONST, 0, cnt, obj.TYPE_REG, ppc64.REGTMP, 0) 87 p = appendpp(p, ppc64.AADD, obj.TYPE_REG, ppc64.REGTMP, 0, obj.TYPE_REG, ppc64.REGRT2, 0) 88 p.Reg = ppc64.REGRT1 89 p = appendpp(p, ppc64.AMOVDU, obj.TYPE_REG, ppc64.REGZERO, 0, obj.TYPE_MEM, ppc64.REGRT1, int64(gc.Widthptr)) 90 p1 := p 91 p = appendpp(p, ppc64.ACMP, obj.TYPE_REG, ppc64.REGRT1, 0, obj.TYPE_REG, ppc64.REGRT2, 0) 92 p = appendpp(p, ppc64.ABNE, obj.TYPE_NONE, 0, 0, obj.TYPE_BRANCH, 0, 0) 93 gc.Patch(p, p1) 94 } 95 96 return p 97 } 98 99 func appendpp(p *obj.Prog, as int, ftype int, freg int, foffset int64, ttype int, treg int, toffset int64) *obj.Prog { 100 q := gc.Ctxt.NewProg() 101 gc.Clearp(q) 102 q.As = int16(as) 103 q.Lineno = p.Lineno 104 q.From.Type = int16(ftype) 105 q.From.Reg = int16(freg) 106 q.From.Offset = foffset 107 q.To.Type = int16(ttype) 108 q.To.Reg = int16(treg) 109 q.To.Offset = toffset 110 q.Link = p.Link 111 p.Link = q 112 return q 113 } 114 115 func ginsnop() { 116 var reg gc.Node 117 gc.Nodreg(®, gc.Types[gc.TINT], ppc64.REG_R0) 118 gins(ppc64.AOR, ®, ®) 119 } 120 121 var panicdiv *gc.Node 122 123 /* 124 * generate division. 125 * generates one of: 126 * res = nl / nr 127 * res = nl % nr 128 * according to op. 129 */ 130 func dodiv(op int, nl *gc.Node, nr *gc.Node, res *gc.Node) { 131 // Have to be careful about handling 132 // most negative int divided by -1 correctly. 133 // The hardware will generate undefined result. 134 // Also need to explicitly trap on division on zero, 135 // the hardware will silently generate undefined result. 136 // DIVW will leave unpredicable result in higher 32-bit, 137 // so always use DIVD/DIVDU. 138 t := nl.Type 139 140 t0 := t 141 check := 0 142 if gc.Issigned[t.Etype] { 143 check = 1 144 if gc.Isconst(nl, gc.CTINT) && nl.Int() != -(1<<uint64(t.Width*8-1)) { 145 check = 0 146 } else if gc.Isconst(nr, gc.CTINT) && nr.Int() != -1 { 147 check = 0 148 } 149 } 150 151 if t.Width < 8 { 152 if gc.Issigned[t.Etype] { 153 t = gc.Types[gc.TINT64] 154 } else { 155 t = gc.Types[gc.TUINT64] 156 } 157 check = 0 158 } 159 160 a := optoas(gc.ODIV, t) 161 162 var tl gc.Node 163 gc.Regalloc(&tl, t0, nil) 164 var tr gc.Node 165 gc.Regalloc(&tr, t0, nil) 166 if nl.Ullman >= nr.Ullman { 167 gc.Cgen(nl, &tl) 168 gc.Cgen(nr, &tr) 169 } else { 170 gc.Cgen(nr, &tr) 171 gc.Cgen(nl, &tl) 172 } 173 174 if t != t0 { 175 // Convert 176 tl2 := tl 177 178 tr2 := tr 179 tl.Type = t 180 tr.Type = t 181 gmove(&tl2, &tl) 182 gmove(&tr2, &tr) 183 } 184 185 // Handle divide-by-zero panic. 186 p1 := gins(optoas(gc.OCMP, t), &tr, nil) 187 188 p1.To.Type = obj.TYPE_REG 189 p1.To.Reg = ppc64.REGZERO 190 p1 = gc.Gbranch(optoas(gc.ONE, t), nil, +1) 191 if panicdiv == nil { 192 panicdiv = gc.Sysfunc("panicdivide") 193 } 194 gc.Ginscall(panicdiv, -1) 195 gc.Patch(p1, gc.Pc) 196 197 var p2 *obj.Prog 198 if check != 0 { 199 var nm1 gc.Node 200 gc.Nodconst(&nm1, t, -1) 201 gins(optoas(gc.OCMP, t), &tr, &nm1) 202 p1 := gc.Gbranch(optoas(gc.ONE, t), nil, +1) 203 if op == gc.ODIV { 204 // a / (-1) is -a. 205 gins(optoas(gc.OMINUS, t), nil, &tl) 206 207 gmove(&tl, res) 208 } else { 209 // a % (-1) is 0. 210 var nz gc.Node 211 gc.Nodconst(&nz, t, 0) 212 213 gmove(&nz, res) 214 } 215 216 p2 = gc.Gbranch(obj.AJMP, nil, 0) 217 gc.Patch(p1, gc.Pc) 218 } 219 220 p1 = gins(a, &tr, &tl) 221 if op == gc.ODIV { 222 gc.Regfree(&tr) 223 gmove(&tl, res) 224 } else { 225 // A%B = A-(A/B*B) 226 var tm gc.Node 227 gc.Regalloc(&tm, t, nil) 228 229 // patch div to use the 3 register form 230 // TODO(minux): add gins3? 231 p1.Reg = p1.To.Reg 232 233 p1.To.Reg = tm.Reg 234 gins(optoas(gc.OMUL, t), &tr, &tm) 235 gc.Regfree(&tr) 236 gins(optoas(gc.OSUB, t), &tm, &tl) 237 gc.Regfree(&tm) 238 gmove(&tl, res) 239 } 240 241 gc.Regfree(&tl) 242 if check != 0 { 243 gc.Patch(p2, gc.Pc) 244 } 245 } 246 247 /* 248 * generate high multiply: 249 * res = (nl*nr) >> width 250 */ 251 func cgen_hmul(nl *gc.Node, nr *gc.Node, res *gc.Node) { 252 // largest ullman on left. 253 if nl.Ullman < nr.Ullman { 254 tmp := (*gc.Node)(nl) 255 nl = nr 256 nr = tmp 257 } 258 259 t := (*gc.Type)(nl.Type) 260 w := int(int(t.Width * 8)) 261 var n1 gc.Node 262 gc.Cgenr(nl, &n1, res) 263 var n2 gc.Node 264 gc.Cgenr(nr, &n2, nil) 265 switch gc.Simtype[t.Etype] { 266 case gc.TINT8, 267 gc.TINT16, 268 gc.TINT32: 269 gins(optoas(gc.OMUL, t), &n2, &n1) 270 p := (*obj.Prog)(gins(ppc64.ASRAD, nil, &n1)) 271 p.From.Type = obj.TYPE_CONST 272 p.From.Offset = int64(w) 273 274 case gc.TUINT8, 275 gc.TUINT16, 276 gc.TUINT32: 277 gins(optoas(gc.OMUL, t), &n2, &n1) 278 p := (*obj.Prog)(gins(ppc64.ASRD, nil, &n1)) 279 p.From.Type = obj.TYPE_CONST 280 p.From.Offset = int64(w) 281 282 case gc.TINT64, 283 gc.TUINT64: 284 if gc.Issigned[t.Etype] { 285 gins(ppc64.AMULHD, &n2, &n1) 286 } else { 287 gins(ppc64.AMULHDU, &n2, &n1) 288 } 289 290 default: 291 gc.Fatal("cgen_hmul %v", t) 292 } 293 294 gc.Cgen(&n1, res) 295 gc.Regfree(&n1) 296 gc.Regfree(&n2) 297 } 298 299 /* 300 * generate shift according to op, one of: 301 * res = nl << nr 302 * res = nl >> nr 303 */ 304 func cgen_shift(op int, bounded bool, nl *gc.Node, nr *gc.Node, res *gc.Node) { 305 a := int(optoas(op, nl.Type)) 306 307 if nr.Op == gc.OLITERAL { 308 var n1 gc.Node 309 gc.Regalloc(&n1, nl.Type, res) 310 gc.Cgen(nl, &n1) 311 sc := uint64(nr.Int()) 312 if sc >= uint64(nl.Type.Width*8) { 313 // large shift gets 2 shifts by width-1 314 var n3 gc.Node 315 gc.Nodconst(&n3, gc.Types[gc.TUINT32], nl.Type.Width*8-1) 316 317 gins(a, &n3, &n1) 318 gins(a, &n3, &n1) 319 } else { 320 gins(a, nr, &n1) 321 } 322 gmove(&n1, res) 323 gc.Regfree(&n1) 324 return 325 } 326 327 if nl.Ullman >= gc.UINF { 328 var n4 gc.Node 329 gc.Tempname(&n4, nl.Type) 330 gc.Cgen(nl, &n4) 331 nl = &n4 332 } 333 334 if nr.Ullman >= gc.UINF { 335 var n5 gc.Node 336 gc.Tempname(&n5, nr.Type) 337 gc.Cgen(nr, &n5) 338 nr = &n5 339 } 340 341 // Allow either uint32 or uint64 as shift type, 342 // to avoid unnecessary conversion from uint32 to uint64 343 // just to do the comparison. 344 tcount := gc.Types[gc.Simtype[nr.Type.Etype]] 345 346 if tcount.Etype < gc.TUINT32 { 347 tcount = gc.Types[gc.TUINT32] 348 } 349 350 var n1 gc.Node 351 gc.Regalloc(&n1, nr.Type, nil) // to hold the shift type in CX 352 var n3 gc.Node 353 gc.Regalloc(&n3, tcount, &n1) // to clear high bits of CX 354 355 var n2 gc.Node 356 gc.Regalloc(&n2, nl.Type, res) 357 358 if nl.Ullman >= nr.Ullman { 359 gc.Cgen(nl, &n2) 360 gc.Cgen(nr, &n1) 361 gmove(&n1, &n3) 362 } else { 363 gc.Cgen(nr, &n1) 364 gmove(&n1, &n3) 365 gc.Cgen(nl, &n2) 366 } 367 368 gc.Regfree(&n3) 369 370 // test and fix up large shifts 371 if !bounded { 372 gc.Nodconst(&n3, tcount, nl.Type.Width*8) 373 gins(optoas(gc.OCMP, tcount), &n1, &n3) 374 p1 := (*obj.Prog)(gc.Gbranch(optoas(gc.OLT, tcount), nil, +1)) 375 if op == gc.ORSH && gc.Issigned[nl.Type.Etype] { 376 gc.Nodconst(&n3, gc.Types[gc.TUINT32], nl.Type.Width*8-1) 377 gins(a, &n3, &n2) 378 } else { 379 gc.Nodconst(&n3, nl.Type, 0) 380 gmove(&n3, &n2) 381 } 382 383 gc.Patch(p1, gc.Pc) 384 } 385 386 gins(a, &n1, &n2) 387 388 gmove(&n2, res) 389 390 gc.Regfree(&n1) 391 gc.Regfree(&n2) 392 } 393 394 func clearfat(nl *gc.Node) { 395 /* clear a fat object */ 396 if gc.Debug['g'] != 0 { 397 fmt.Printf("clearfat %v (%v, size: %d)\n", nl, nl.Type, nl.Type.Width) 398 } 399 400 w := uint64(uint64(nl.Type.Width)) 401 402 // Avoid taking the address for simple enough types. 403 if gc.Componentgen(nil, nl) { 404 return 405 } 406 407 c := uint64(w % 8) // bytes 408 q := uint64(w / 8) // dwords 409 410 if gc.Reginuse(ppc64.REGRT1) { 411 gc.Fatal("%v in use during clearfat", obj.Rconv(ppc64.REGRT1)) 412 } 413 414 var r0 gc.Node 415 gc.Nodreg(&r0, gc.Types[gc.TUINT64], ppc64.REGZERO) 416 var dst gc.Node 417 gc.Nodreg(&dst, gc.Types[gc.Tptr], ppc64.REGRT1) 418 gc.Regrealloc(&dst) 419 gc.Agen(nl, &dst) 420 421 var boff uint64 422 if q > 128 { 423 p := gins(ppc64.ASUB, nil, &dst) 424 p.From.Type = obj.TYPE_CONST 425 p.From.Offset = 8 426 427 var end gc.Node 428 gc.Regalloc(&end, gc.Types[gc.Tptr], nil) 429 p = gins(ppc64.AMOVD, &dst, &end) 430 p.From.Type = obj.TYPE_ADDR 431 p.From.Offset = int64(q * 8) 432 433 p = gins(ppc64.AMOVDU, &r0, &dst) 434 p.To.Type = obj.TYPE_MEM 435 p.To.Offset = 8 436 pl := (*obj.Prog)(p) 437 438 p = gins(ppc64.ACMP, &dst, &end) 439 gc.Patch(gc.Gbranch(ppc64.ABNE, nil, 0), pl) 440 441 gc.Regfree(&end) 442 443 // The loop leaves R3 on the last zeroed dword 444 boff = 8 445 } else if q >= 4 { 446 p := gins(ppc64.ASUB, nil, &dst) 447 p.From.Type = obj.TYPE_CONST 448 p.From.Offset = 8 449 f := (*gc.Node)(gc.Sysfunc("duffzero")) 450 p = gins(obj.ADUFFZERO, nil, f) 451 gc.Afunclit(&p.To, f) 452 453 // 4 and 128 = magic constants: see ../../runtime/asm_ppc64x.s 454 p.To.Offset = int64(4 * (128 - q)) 455 456 // duffzero leaves R3 on the last zeroed dword 457 boff = 8 458 } else { 459 var p *obj.Prog 460 for t := uint64(0); t < q; t++ { 461 p = gins(ppc64.AMOVD, &r0, &dst) 462 p.To.Type = obj.TYPE_MEM 463 p.To.Offset = int64(8 * t) 464 } 465 466 boff = 8 * q 467 } 468 469 var p *obj.Prog 470 for t := uint64(0); t < c; t++ { 471 p = gins(ppc64.AMOVB, &r0, &dst) 472 p.To.Type = obj.TYPE_MEM 473 p.To.Offset = int64(t + boff) 474 } 475 476 gc.Regfree(&dst) 477 } 478 479 // Called after regopt and peep have run. 480 // Expand CHECKNIL pseudo-op into actual nil pointer check. 481 func expandchecks(firstp *obj.Prog) { 482 var p1 *obj.Prog 483 var p2 *obj.Prog 484 485 for p := (*obj.Prog)(firstp); p != nil; p = p.Link { 486 if gc.Debug_checknil != 0 && gc.Ctxt.Debugvlog != 0 { 487 fmt.Printf("expandchecks: %v\n", p) 488 } 489 if p.As != obj.ACHECKNIL { 490 continue 491 } 492 if gc.Debug_checknil != 0 && p.Lineno > 1 { // p->lineno==1 in generated wrappers 493 gc.Warnl(int(p.Lineno), "generated nil check") 494 } 495 if p.From.Type != obj.TYPE_REG { 496 gc.Fatal("invalid nil check %v\n", p) 497 } 498 499 /* 500 // check is 501 // TD $4, R0, arg (R0 is always zero) 502 // eqv. to: 503 // tdeq r0, arg 504 // NOTE: this needs special runtime support to make SIGTRAP recoverable. 505 reg = p->from.reg; 506 p->as = ATD; 507 p->from = p->to = p->from3 = zprog.from; 508 p->from.type = TYPE_CONST; 509 p->from.offset = 4; 510 p->from.reg = 0; 511 p->reg = REGZERO; 512 p->to.type = TYPE_REG; 513 p->to.reg = reg; 514 */ 515 // check is 516 // CMP arg, R0 517 // BNE 2(PC) [likely] 518 // MOVD R0, 0(R0) 519 p1 = gc.Ctxt.NewProg() 520 521 p2 = gc.Ctxt.NewProg() 522 gc.Clearp(p1) 523 gc.Clearp(p2) 524 p1.Link = p2 525 p2.Link = p.Link 526 p.Link = p1 527 p1.Lineno = p.Lineno 528 p2.Lineno = p.Lineno 529 p1.Pc = 9999 530 p2.Pc = 9999 531 p.As = ppc64.ACMP 532 p.To.Type = obj.TYPE_REG 533 p.To.Reg = ppc64.REGZERO 534 p1.As = ppc64.ABNE 535 536 //p1->from.type = TYPE_CONST; 537 //p1->from.offset = 1; // likely 538 p1.To.Type = obj.TYPE_BRANCH 539 540 p1.To.Val = p2.Link 541 542 // crash by write to memory address 0. 543 p2.As = ppc64.AMOVD 544 545 p2.From.Type = obj.TYPE_REG 546 p2.From.Reg = ppc64.REGZERO 547 p2.To.Type = obj.TYPE_MEM 548 p2.To.Reg = ppc64.REGZERO 549 p2.To.Offset = 0 550 } 551 } 552 553 // res = runtime.getg() 554 func getg(res *gc.Node) { 555 var n1 gc.Node 556 gc.Nodreg(&n1, res.Type, ppc64.REGG) 557 gmove(&n1, res) 558 }