github.com/jonasi/go@v0.0.0-20150930005915-e78e654c1de0/src/cmd/compile/internal/ppc64/ggen.go (about) 1 // Copyright 2009 The Go Authors. All rights reserved. 2 // Use of this source code is governed by a BSD-style 3 // license that can be found in the LICENSE file. 4 5 package ppc64 6 7 import ( 8 "cmd/compile/internal/gc" 9 "cmd/internal/obj" 10 "cmd/internal/obj/ppc64" 11 "fmt" 12 ) 13 14 func defframe(ptxt *obj.Prog) { 15 var n *gc.Node 16 17 // fill in argument size, stack size 18 ptxt.To.Type = obj.TYPE_TEXTSIZE 19 20 ptxt.To.Val = int32(gc.Rnd(gc.Curfn.Type.Argwid, int64(gc.Widthptr))) 21 frame := uint32(gc.Rnd(gc.Stksize+gc.Maxarg, int64(gc.Widthreg))) 22 ptxt.To.Offset = int64(frame) 23 24 // insert code to zero ambiguously live variables 25 // so that the garbage collector only sees initialized values 26 // when it looks for pointers. 27 p := ptxt 28 29 hi := int64(0) 30 lo := hi 31 32 // iterate through declarations - they are sorted in decreasing xoffset order. 33 for l := gc.Curfn.Func.Dcl; l != nil; l = l.Next { 34 n = l.N 35 if !n.Name.Needzero { 36 continue 37 } 38 if n.Class != gc.PAUTO { 39 gc.Fatalf("needzero class %d", n.Class) 40 } 41 if n.Type.Width%int64(gc.Widthptr) != 0 || n.Xoffset%int64(gc.Widthptr) != 0 || n.Type.Width == 0 { 42 gc.Fatalf("var %v has size %d offset %d", gc.Nconv(n, obj.FmtLong), int(n.Type.Width), int(n.Xoffset)) 43 } 44 45 if lo != hi && n.Xoffset+n.Type.Width >= lo-int64(2*gc.Widthreg) { 46 // merge with range we already have 47 lo = n.Xoffset 48 49 continue 50 } 51 52 // zero old range 53 p = zerorange(p, int64(frame), lo, hi) 54 55 // set new range 56 hi = n.Xoffset + n.Type.Width 57 58 lo = n.Xoffset 59 } 60 61 // zero final range 62 zerorange(p, int64(frame), lo, hi) 63 } 64 65 func zerorange(p *obj.Prog, frame int64, lo int64, hi int64) *obj.Prog { 66 cnt := hi - lo 67 if cnt == 0 { 68 return p 69 } 70 if cnt < int64(4*gc.Widthptr) { 71 for i := int64(0); i < cnt; i += int64(gc.Widthptr) { 72 p = appendpp(p, ppc64.AMOVD, obj.TYPE_REG, ppc64.REGZERO, 0, obj.TYPE_MEM, ppc64.REGSP, 8+frame+lo+i) 73 } 74 // TODO(dfc): https://golang.org/issue/12108 75 // If DUFFZERO is used inside a tail call (see genwrapper) it will 76 // overwrite the link register. 77 } else if false && cnt <= int64(128*gc.Widthptr) { 78 p = appendpp(p, ppc64.AADD, obj.TYPE_CONST, 0, 8+frame+lo-8, obj.TYPE_REG, ppc64.REGRT1, 0) 79 p.Reg = ppc64.REGSP 80 p = appendpp(p, obj.ADUFFZERO, obj.TYPE_NONE, 0, 0, obj.TYPE_MEM, 0, 0) 81 f := gc.Sysfunc("duffzero") 82 gc.Naddr(&p.To, f) 83 gc.Afunclit(&p.To, f) 84 p.To.Offset = 4 * (128 - cnt/int64(gc.Widthptr)) 85 } else { 86 p = appendpp(p, ppc64.AMOVD, obj.TYPE_CONST, 0, 8+frame+lo-8, obj.TYPE_REG, ppc64.REGTMP, 0) 87 p = appendpp(p, ppc64.AADD, obj.TYPE_REG, ppc64.REGTMP, 0, obj.TYPE_REG, ppc64.REGRT1, 0) 88 p.Reg = ppc64.REGSP 89 p = appendpp(p, ppc64.AMOVD, obj.TYPE_CONST, 0, cnt, obj.TYPE_REG, ppc64.REGTMP, 0) 90 p = appendpp(p, ppc64.AADD, obj.TYPE_REG, ppc64.REGTMP, 0, obj.TYPE_REG, ppc64.REGRT2, 0) 91 p.Reg = ppc64.REGRT1 92 p = appendpp(p, ppc64.AMOVDU, obj.TYPE_REG, ppc64.REGZERO, 0, obj.TYPE_MEM, ppc64.REGRT1, int64(gc.Widthptr)) 93 p1 := p 94 p = appendpp(p, ppc64.ACMP, obj.TYPE_REG, ppc64.REGRT1, 0, obj.TYPE_REG, ppc64.REGRT2, 0) 95 p = appendpp(p, ppc64.ABNE, obj.TYPE_NONE, 0, 0, obj.TYPE_BRANCH, 0, 0) 96 gc.Patch(p, p1) 97 } 98 99 return p 100 } 101 102 func appendpp(p *obj.Prog, as int, ftype int, freg int, foffset int64, ttype int, treg int, toffset int64) *obj.Prog { 103 q := gc.Ctxt.NewProg() 104 gc.Clearp(q) 105 q.As = int16(as) 106 q.Lineno = p.Lineno 107 q.From.Type = int16(ftype) 108 q.From.Reg = int16(freg) 109 q.From.Offset = foffset 110 q.To.Type = int16(ttype) 111 q.To.Reg = int16(treg) 112 q.To.Offset = toffset 113 q.Link = p.Link 114 p.Link = q 115 return q 116 } 117 118 func ginsnop() { 119 var reg gc.Node 120 gc.Nodreg(®, gc.Types[gc.TINT], ppc64.REG_R0) 121 gins(ppc64.AOR, ®, ®) 122 } 123 124 var panicdiv *gc.Node 125 126 /* 127 * generate division. 128 * generates one of: 129 * res = nl / nr 130 * res = nl % nr 131 * according to op. 132 */ 133 func dodiv(op int, nl *gc.Node, nr *gc.Node, res *gc.Node) { 134 // Have to be careful about handling 135 // most negative int divided by -1 correctly. 136 // The hardware will generate undefined result. 137 // Also need to explicitly trap on division on zero, 138 // the hardware will silently generate undefined result. 139 // DIVW will leave unpredicable result in higher 32-bit, 140 // so always use DIVD/DIVDU. 141 t := nl.Type 142 143 t0 := t 144 check := false 145 if gc.Issigned[t.Etype] { 146 check = true 147 if gc.Isconst(nl, gc.CTINT) && nl.Int() != -(1<<uint64(t.Width*8-1)) { 148 check = false 149 } else if gc.Isconst(nr, gc.CTINT) && nr.Int() != -1 { 150 check = false 151 } 152 } 153 154 if t.Width < 8 { 155 if gc.Issigned[t.Etype] { 156 t = gc.Types[gc.TINT64] 157 } else { 158 t = gc.Types[gc.TUINT64] 159 } 160 check = false 161 } 162 163 a := optoas(gc.ODIV, t) 164 165 var tl gc.Node 166 gc.Regalloc(&tl, t0, nil) 167 var tr gc.Node 168 gc.Regalloc(&tr, t0, nil) 169 if nl.Ullman >= nr.Ullman { 170 gc.Cgen(nl, &tl) 171 gc.Cgen(nr, &tr) 172 } else { 173 gc.Cgen(nr, &tr) 174 gc.Cgen(nl, &tl) 175 } 176 177 if t != t0 { 178 // Convert 179 tl2 := tl 180 181 tr2 := tr 182 tl.Type = t 183 tr.Type = t 184 gmove(&tl2, &tl) 185 gmove(&tr2, &tr) 186 } 187 188 // Handle divide-by-zero panic. 189 p1 := gins(optoas(gc.OCMP, t), &tr, nil) 190 191 p1.To.Type = obj.TYPE_REG 192 p1.To.Reg = ppc64.REGZERO 193 p1 = gc.Gbranch(optoas(gc.ONE, t), nil, +1) 194 if panicdiv == nil { 195 panicdiv = gc.Sysfunc("panicdivide") 196 } 197 gc.Ginscall(panicdiv, -1) 198 gc.Patch(p1, gc.Pc) 199 200 var p2 *obj.Prog 201 if check { 202 var nm1 gc.Node 203 gc.Nodconst(&nm1, t, -1) 204 gins(optoas(gc.OCMP, t), &tr, &nm1) 205 p1 := gc.Gbranch(optoas(gc.ONE, t), nil, +1) 206 if op == gc.ODIV { 207 // a / (-1) is -a. 208 gins(optoas(gc.OMINUS, t), nil, &tl) 209 210 gmove(&tl, res) 211 } else { 212 // a % (-1) is 0. 213 var nz gc.Node 214 gc.Nodconst(&nz, t, 0) 215 216 gmove(&nz, res) 217 } 218 219 p2 = gc.Gbranch(obj.AJMP, nil, 0) 220 gc.Patch(p1, gc.Pc) 221 } 222 223 p1 = gins(a, &tr, &tl) 224 if op == gc.ODIV { 225 gc.Regfree(&tr) 226 gmove(&tl, res) 227 } else { 228 // A%B = A-(A/B*B) 229 var tm gc.Node 230 gc.Regalloc(&tm, t, nil) 231 232 // patch div to use the 3 register form 233 // TODO(minux): add gins3? 234 p1.Reg = p1.To.Reg 235 236 p1.To.Reg = tm.Reg 237 gins(optoas(gc.OMUL, t), &tr, &tm) 238 gc.Regfree(&tr) 239 gins(optoas(gc.OSUB, t), &tm, &tl) 240 gc.Regfree(&tm) 241 gmove(&tl, res) 242 } 243 244 gc.Regfree(&tl) 245 if check { 246 gc.Patch(p2, gc.Pc) 247 } 248 } 249 250 /* 251 * generate high multiply: 252 * res = (nl*nr) >> width 253 */ 254 func cgen_hmul(nl *gc.Node, nr *gc.Node, res *gc.Node) { 255 // largest ullman on left. 256 if nl.Ullman < nr.Ullman { 257 nl, nr = nr, nl 258 } 259 260 t := (*gc.Type)(nl.Type) 261 w := int(int(t.Width * 8)) 262 var n1 gc.Node 263 gc.Cgenr(nl, &n1, res) 264 var n2 gc.Node 265 gc.Cgenr(nr, &n2, nil) 266 switch gc.Simtype[t.Etype] { 267 case gc.TINT8, 268 gc.TINT16, 269 gc.TINT32: 270 gins(optoas(gc.OMUL, t), &n2, &n1) 271 p := (*obj.Prog)(gins(ppc64.ASRAD, nil, &n1)) 272 p.From.Type = obj.TYPE_CONST 273 p.From.Offset = int64(w) 274 275 case gc.TUINT8, 276 gc.TUINT16, 277 gc.TUINT32: 278 gins(optoas(gc.OMUL, t), &n2, &n1) 279 p := (*obj.Prog)(gins(ppc64.ASRD, nil, &n1)) 280 p.From.Type = obj.TYPE_CONST 281 p.From.Offset = int64(w) 282 283 case gc.TINT64, 284 gc.TUINT64: 285 if gc.Issigned[t.Etype] { 286 gins(ppc64.AMULHD, &n2, &n1) 287 } else { 288 gins(ppc64.AMULHDU, &n2, &n1) 289 } 290 291 default: 292 gc.Fatalf("cgen_hmul %v", t) 293 } 294 295 gc.Cgen(&n1, res) 296 gc.Regfree(&n1) 297 gc.Regfree(&n2) 298 } 299 300 /* 301 * generate shift according to op, one of: 302 * res = nl << nr 303 * res = nl >> nr 304 */ 305 func cgen_shift(op int, bounded bool, nl *gc.Node, nr *gc.Node, res *gc.Node) { 306 a := int(optoas(op, nl.Type)) 307 308 if nr.Op == gc.OLITERAL { 309 var n1 gc.Node 310 gc.Regalloc(&n1, nl.Type, res) 311 gc.Cgen(nl, &n1) 312 sc := uint64(nr.Int()) 313 if sc >= uint64(nl.Type.Width*8) { 314 // large shift gets 2 shifts by width-1 315 var n3 gc.Node 316 gc.Nodconst(&n3, gc.Types[gc.TUINT32], nl.Type.Width*8-1) 317 318 gins(a, &n3, &n1) 319 gins(a, &n3, &n1) 320 } else { 321 gins(a, nr, &n1) 322 } 323 gmove(&n1, res) 324 gc.Regfree(&n1) 325 return 326 } 327 328 if nl.Ullman >= gc.UINF { 329 var n4 gc.Node 330 gc.Tempname(&n4, nl.Type) 331 gc.Cgen(nl, &n4) 332 nl = &n4 333 } 334 335 if nr.Ullman >= gc.UINF { 336 var n5 gc.Node 337 gc.Tempname(&n5, nr.Type) 338 gc.Cgen(nr, &n5) 339 nr = &n5 340 } 341 342 // Allow either uint32 or uint64 as shift type, 343 // to avoid unnecessary conversion from uint32 to uint64 344 // just to do the comparison. 345 tcount := gc.Types[gc.Simtype[nr.Type.Etype]] 346 347 if tcount.Etype < gc.TUINT32 { 348 tcount = gc.Types[gc.TUINT32] 349 } 350 351 var n1 gc.Node 352 gc.Regalloc(&n1, nr.Type, nil) // to hold the shift type in CX 353 var n3 gc.Node 354 gc.Regalloc(&n3, tcount, &n1) // to clear high bits of CX 355 356 var n2 gc.Node 357 gc.Regalloc(&n2, nl.Type, res) 358 359 if nl.Ullman >= nr.Ullman { 360 gc.Cgen(nl, &n2) 361 gc.Cgen(nr, &n1) 362 gmove(&n1, &n3) 363 } else { 364 gc.Cgen(nr, &n1) 365 gmove(&n1, &n3) 366 gc.Cgen(nl, &n2) 367 } 368 369 gc.Regfree(&n3) 370 371 // test and fix up large shifts 372 if !bounded { 373 gc.Nodconst(&n3, tcount, nl.Type.Width*8) 374 gins(optoas(gc.OCMP, tcount), &n1, &n3) 375 p1 := (*obj.Prog)(gc.Gbranch(optoas(gc.OLT, tcount), nil, +1)) 376 if op == gc.ORSH && gc.Issigned[nl.Type.Etype] { 377 gc.Nodconst(&n3, gc.Types[gc.TUINT32], nl.Type.Width*8-1) 378 gins(a, &n3, &n2) 379 } else { 380 gc.Nodconst(&n3, nl.Type, 0) 381 gmove(&n3, &n2) 382 } 383 384 gc.Patch(p1, gc.Pc) 385 } 386 387 gins(a, &n1, &n2) 388 389 gmove(&n2, res) 390 391 gc.Regfree(&n1) 392 gc.Regfree(&n2) 393 } 394 395 func clearfat(nl *gc.Node) { 396 /* clear a fat object */ 397 if gc.Debug['g'] != 0 { 398 fmt.Printf("clearfat %v (%v, size: %d)\n", nl, nl.Type, nl.Type.Width) 399 } 400 401 w := uint64(uint64(nl.Type.Width)) 402 403 // Avoid taking the address for simple enough types. 404 if gc.Componentgen(nil, nl) { 405 return 406 } 407 408 c := uint64(w % 8) // bytes 409 q := uint64(w / 8) // dwords 410 411 if gc.Reginuse(ppc64.REGRT1) { 412 gc.Fatalf("%v in use during clearfat", obj.Rconv(ppc64.REGRT1)) 413 } 414 415 var r0 gc.Node 416 gc.Nodreg(&r0, gc.Types[gc.TUINT64], ppc64.REGZERO) 417 var dst gc.Node 418 gc.Nodreg(&dst, gc.Types[gc.Tptr], ppc64.REGRT1) 419 gc.Regrealloc(&dst) 420 gc.Agen(nl, &dst) 421 422 var boff uint64 423 if q > 128 { 424 p := gins(ppc64.ASUB, nil, &dst) 425 p.From.Type = obj.TYPE_CONST 426 p.From.Offset = 8 427 428 var end gc.Node 429 gc.Regalloc(&end, gc.Types[gc.Tptr], nil) 430 p = gins(ppc64.AMOVD, &dst, &end) 431 p.From.Type = obj.TYPE_ADDR 432 p.From.Offset = int64(q * 8) 433 434 p = gins(ppc64.AMOVDU, &r0, &dst) 435 p.To.Type = obj.TYPE_MEM 436 p.To.Offset = 8 437 pl := (*obj.Prog)(p) 438 439 p = gins(ppc64.ACMP, &dst, &end) 440 gc.Patch(gc.Gbranch(ppc64.ABNE, nil, 0), pl) 441 442 gc.Regfree(&end) 443 444 // The loop leaves R3 on the last zeroed dword 445 boff = 8 446 // TODO(dfc): https://golang.org/issue/12108 447 // If DUFFZERO is used inside a tail call (see genwrapper) it will 448 // overwrite the link register. 449 } else if false && q >= 4 { 450 p := gins(ppc64.ASUB, nil, &dst) 451 p.From.Type = obj.TYPE_CONST 452 p.From.Offset = 8 453 f := (*gc.Node)(gc.Sysfunc("duffzero")) 454 p = gins(obj.ADUFFZERO, nil, f) 455 gc.Afunclit(&p.To, f) 456 457 // 4 and 128 = magic constants: see ../../runtime/asm_ppc64x.s 458 p.To.Offset = int64(4 * (128 - q)) 459 460 // duffzero leaves R3 on the last zeroed dword 461 boff = 8 462 } else { 463 var p *obj.Prog 464 for t := uint64(0); t < q; t++ { 465 p = gins(ppc64.AMOVD, &r0, &dst) 466 p.To.Type = obj.TYPE_MEM 467 p.To.Offset = int64(8 * t) 468 } 469 470 boff = 8 * q 471 } 472 473 var p *obj.Prog 474 for t := uint64(0); t < c; t++ { 475 p = gins(ppc64.AMOVB, &r0, &dst) 476 p.To.Type = obj.TYPE_MEM 477 p.To.Offset = int64(t + boff) 478 } 479 480 gc.Regfree(&dst) 481 } 482 483 // Called after regopt and peep have run. 484 // Expand CHECKNIL pseudo-op into actual nil pointer check. 485 func expandchecks(firstp *obj.Prog) { 486 var p1 *obj.Prog 487 var p2 *obj.Prog 488 489 for p := (*obj.Prog)(firstp); p != nil; p = p.Link { 490 if gc.Debug_checknil != 0 && gc.Ctxt.Debugvlog != 0 { 491 fmt.Printf("expandchecks: %v\n", p) 492 } 493 if p.As != obj.ACHECKNIL { 494 continue 495 } 496 if gc.Debug_checknil != 0 && p.Lineno > 1 { // p->lineno==1 in generated wrappers 497 gc.Warnl(int(p.Lineno), "generated nil check") 498 } 499 if p.From.Type != obj.TYPE_REG { 500 gc.Fatalf("invalid nil check %v\n", p) 501 } 502 503 /* 504 // check is 505 // TD $4, R0, arg (R0 is always zero) 506 // eqv. to: 507 // tdeq r0, arg 508 // NOTE: this needs special runtime support to make SIGTRAP recoverable. 509 reg = p->from.reg; 510 p->as = ATD; 511 p->from = p->to = p->from3 = zprog.from; 512 p->from.type = TYPE_CONST; 513 p->from.offset = 4; 514 p->from.reg = 0; 515 p->reg = REGZERO; 516 p->to.type = TYPE_REG; 517 p->to.reg = reg; 518 */ 519 // check is 520 // CMP arg, R0 521 // BNE 2(PC) [likely] 522 // MOVD R0, 0(R0) 523 p1 = gc.Ctxt.NewProg() 524 525 p2 = gc.Ctxt.NewProg() 526 gc.Clearp(p1) 527 gc.Clearp(p2) 528 p1.Link = p2 529 p2.Link = p.Link 530 p.Link = p1 531 p1.Lineno = p.Lineno 532 p2.Lineno = p.Lineno 533 p1.Pc = 9999 534 p2.Pc = 9999 535 p.As = ppc64.ACMP 536 p.To.Type = obj.TYPE_REG 537 p.To.Reg = ppc64.REGZERO 538 p1.As = ppc64.ABNE 539 540 //p1->from.type = TYPE_CONST; 541 //p1->from.offset = 1; // likely 542 p1.To.Type = obj.TYPE_BRANCH 543 544 p1.To.Val = p2.Link 545 546 // crash by write to memory address 0. 547 p2.As = ppc64.AMOVD 548 549 p2.From.Type = obj.TYPE_REG 550 p2.From.Reg = ppc64.REGZERO 551 p2.To.Type = obj.TYPE_MEM 552 p2.To.Reg = ppc64.REGZERO 553 p2.To.Offset = 0 554 } 555 } 556 557 // res = runtime.getg() 558 func getg(res *gc.Node) { 559 var n1 gc.Node 560 gc.Nodreg(&n1, res.Type, ppc64.REGG) 561 gmove(&n1, res) 562 }