1 // Derived from Inferno utils/8c/txt.c
2 // http://code.google.com/p/inferno-os/source/browse/utils/8c/txt.c
4 // Copyright © 1994-1999 Lucent Technologies Inc. All rights reserved.
5 // Portions Copyright © 1995-1997 C H Forsyth (forsyth@terzarima.net)
6 // Portions Copyright © 1997-1999 Vita Nuova Limited
7 // Portions Copyright © 2000-2007 Vita Nuova Holdings Limited (www.vitanuova.com)
8 // Portions Copyright © 2004,2006 Bruce Ellis
9 // Portions Copyright © 2005-2007 C H Forsyth (forsyth@terzarima.net)
10 // Revisions Copyright © 2000-2007 Lucent Technologies Inc. and others
11 // Portions Copyright © 2009 The Go Authors. All rights reserved.
13 // Permission is hereby granted, free of charge, to any person obtaining a copy
14 // of this software and associated documentation files (the "Software"), to deal
15 // in the Software without restriction, including without limitation the rights
16 // to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
17 // copies of the Software, and to permit persons to whom the Software is
18 // furnished to do so, subject to the following conditions:
20 // The above copyright notice and this permission notice shall be included in
21 // all copies or substantial portions of the Software.
23 // THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
24 // IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
25 // FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
26 // AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
27 // LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
28 // OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN
34 "cmd/compile/internal/big"
35 "cmd/compile/internal/gc"
37 "cmd/internal/obj/x86"
41 // TODO(rsc): Can make this bigger if we move
42 // the text segment up higher in 8l for all GOOS.
43 // At the same time, can raise StackBig in ../../runtime/stack.h.
44 var unmappedzero uint32 = 4096
54 * return Axxx for Oxxx on type t.
56 func optoas(op gc.Op, t *gc.Type) obj.As {
58 gc.Fatalf("optoas: t is nil")
61 // avoid constant conversions in switches below
63 OMINUS_ = uint32(gc.OMINUS) << 16
64 OLSH_ = uint32(gc.OLSH) << 16
65 ORSH_ = uint32(gc.ORSH) << 16
66 OADD_ = uint32(gc.OADD) << 16
67 OSUB_ = uint32(gc.OSUB) << 16
68 OMUL_ = uint32(gc.OMUL) << 16
69 ODIV_ = uint32(gc.ODIV) << 16
70 OMOD_ = uint32(gc.OMOD) << 16
71 OOR_ = uint32(gc.OOR) << 16
72 OAND_ = uint32(gc.OAND) << 16
73 OXOR_ = uint32(gc.OXOR) << 16
74 OEQ_ = uint32(gc.OEQ) << 16
75 ONE_ = uint32(gc.ONE) << 16
76 OLT_ = uint32(gc.OLT) << 16
77 OLE_ = uint32(gc.OLE) << 16
78 OGE_ = uint32(gc.OGE) << 16
79 OGT_ = uint32(gc.OGT) << 16
80 OCMP_ = uint32(gc.OCMP) << 16
81 OAS_ = uint32(gc.OAS) << 16
82 OHMUL_ = uint32(gc.OHMUL) << 16
83 OADDR_ = uint32(gc.OADDR) << 16
84 OINC_ = uint32(gc.OINC) << 16
85 ODEC_ = uint32(gc.ODEC) << 16
86 OLROT_ = uint32(gc.OLROT) << 16
87 OEXTEND_ = uint32(gc.OEXTEND) << 16
88 OCOM_ = uint32(gc.OCOM) << 16
92 switch uint32(op)<<16 | uint32(gc.Simtype[t.Etype]) {
94 gc.Fatalf("optoas: no entry %v-%v", op, t)
96 case OADDR_ | gc.TPTR32:
114 case ONE_ | gc.TBOOL,
129 case OLT_ | gc.TINT8,
135 case OLT_ | gc.TUINT8,
141 case OLE_ | gc.TINT8,
147 case OLE_ | gc.TUINT8,
153 case OGT_ | gc.TINT8,
159 case OGT_ | gc.TUINT8,
167 case OGE_ | gc.TINT8,
173 case OGE_ | gc.TUINT8,
181 case OCMP_ | gc.TBOOL,
186 case OCMP_ | gc.TINT16,
190 case OCMP_ | gc.TINT32,
195 case OAS_ | gc.TBOOL,
200 case OAS_ | gc.TINT16,
204 case OAS_ | gc.TINT32,
209 case OAS_ | gc.TFLOAT32:
212 case OAS_ | gc.TFLOAT64:
215 case OADD_ | gc.TINT8,
219 case OADD_ | gc.TINT16,
223 case OADD_ | gc.TINT32,
228 case OSUB_ | gc.TINT8,
232 case OSUB_ | gc.TINT16,
236 case OSUB_ | gc.TINT32,
241 case OINC_ | gc.TINT8,
245 case OINC_ | gc.TINT16,
249 case OINC_ | gc.TINT32,
254 case ODEC_ | gc.TINT8,
258 case ODEC_ | gc.TINT16,
262 case ODEC_ | gc.TINT32,
267 case OCOM_ | gc.TINT8,
271 case OCOM_ | gc.TINT16,
275 case OCOM_ | gc.TINT32,
280 case OMINUS_ | gc.TINT8,
284 case OMINUS_ | gc.TINT16,
285 OMINUS_ | gc.TUINT16:
288 case OMINUS_ | gc.TINT32,
289 OMINUS_ | gc.TUINT32,
293 case OAND_ | gc.TINT8,
297 case OAND_ | gc.TINT16,
301 case OAND_ | gc.TINT32,
306 case OOR_ | gc.TINT8,
310 case OOR_ | gc.TINT16,
314 case OOR_ | gc.TINT32,
319 case OXOR_ | gc.TINT8,
323 case OXOR_ | gc.TINT16,
327 case OXOR_ | gc.TINT32,
332 case OLROT_ | gc.TINT8,
336 case OLROT_ | gc.TINT16,
340 case OLROT_ | gc.TINT32,
345 case OLSH_ | gc.TINT8,
349 case OLSH_ | gc.TINT16,
353 case OLSH_ | gc.TINT32,
358 case ORSH_ | gc.TUINT8:
361 case ORSH_ | gc.TUINT16:
364 case ORSH_ | gc.TUINT32,
368 case ORSH_ | gc.TINT8:
371 case ORSH_ | gc.TINT16:
374 case ORSH_ | gc.TINT32:
377 case OHMUL_ | gc.TINT8,
382 case OHMUL_ | gc.TINT16,
387 case OHMUL_ | gc.TINT32,
393 case OHMUL_ | gc.TUINT8:
396 case OHMUL_ | gc.TUINT16:
399 case OHMUL_ | gc.TUINT32,
403 case ODIV_ | gc.TINT8,
407 case ODIV_ | gc.TUINT8,
411 case ODIV_ | gc.TINT16,
415 case ODIV_ | gc.TUINT16,
419 case ODIV_ | gc.TINT32,
423 case ODIV_ | gc.TUINT32,
429 case OEXTEND_ | gc.TINT16:
432 case OEXTEND_ | gc.TINT32:
439 func foptoas(op gc.Op, t *gc.Type, flg int) obj.As {
441 et := gc.Simtype[t.Etype]
443 // avoid constant conversions in switches below
445 OCMP_ = uint32(gc.OCMP) << 16
446 OAS_ = uint32(gc.OAS) << 16
447 OADD_ = uint32(gc.OADD) << 16
448 OSUB_ = uint32(gc.OSUB) << 16
449 OMUL_ = uint32(gc.OMUL) << 16
450 ODIV_ = uint32(gc.ODIV) << 16
451 OMINUS_ = uint32(gc.OMINUS) << 16
454 if !gc.Thearch.Use387 {
455 switch uint32(op)<<16 | uint32(et) {
457 gc.Fatalf("foptoas-sse: no entry %v-%v", op, t)
459 case OCMP_ | gc.TFLOAT32:
462 case OCMP_ | gc.TFLOAT64:
465 case OAS_ | gc.TFLOAT32:
468 case OAS_ | gc.TFLOAT64:
471 case OADD_ | gc.TFLOAT32:
474 case OADD_ | gc.TFLOAT64:
477 case OSUB_ | gc.TFLOAT32:
480 case OSUB_ | gc.TFLOAT64:
483 case OMUL_ | gc.TFLOAT32:
486 case OMUL_ | gc.TFLOAT64:
489 case ODIV_ | gc.TFLOAT32:
492 case ODIV_ | gc.TFLOAT64:
499 // If we need Fpop, it means we're working on
500 // two different floating-point registers, not memory.
501 // There the instruction only has a float64 form.
506 // clear Frev if unneeded
513 switch uint32(op)<<16 | (uint32(et)<<8 | uint32(flg)) {
514 case OADD_ | (gc.TFLOAT32<<8 | 0):
517 case OADD_ | (gc.TFLOAT64<<8 | 0):
520 case OADD_ | (gc.TFLOAT64<<8 | Fpop):
523 case OSUB_ | (gc.TFLOAT32<<8 | 0):
526 case OSUB_ | (gc.TFLOAT32<<8 | Frev):
529 case OSUB_ | (gc.TFLOAT64<<8 | 0):
532 case OSUB_ | (gc.TFLOAT64<<8 | Frev):
535 case OSUB_ | (gc.TFLOAT64<<8 | Fpop):
538 case OSUB_ | (gc.TFLOAT64<<8 | (Fpop | Frev)):
541 case OMUL_ | (gc.TFLOAT32<<8 | 0):
544 case OMUL_ | (gc.TFLOAT64<<8 | 0):
547 case OMUL_ | (gc.TFLOAT64<<8 | Fpop):
550 case ODIV_ | (gc.TFLOAT32<<8 | 0):
553 case ODIV_ | (gc.TFLOAT32<<8 | Frev):
556 case ODIV_ | (gc.TFLOAT64<<8 | 0):
559 case ODIV_ | (gc.TFLOAT64<<8 | Frev):
562 case ODIV_ | (gc.TFLOAT64<<8 | Fpop):
565 case ODIV_ | (gc.TFLOAT64<<8 | (Fpop | Frev)):
568 case OCMP_ | (gc.TFLOAT32<<8 | 0):
571 case OCMP_ | (gc.TFLOAT32<<8 | Fpop):
574 case OCMP_ | (gc.TFLOAT64<<8 | 0):
577 case OCMP_ | (gc.TFLOAT64<<8 | Fpop):
580 case OCMP_ | (gc.TFLOAT64<<8 | Fpop2):
583 case OMINUS_ | (gc.TFLOAT32<<8 | 0):
586 case OMINUS_ | (gc.TFLOAT64<<8 | 0):
590 gc.Fatalf("foptoas %v %v %#x", op, t, flg)
595 // REG_DI, // for movstring
596 // REG_SI, // for movstring
598 x86.REG_AX, // for divide
599 x86.REG_CX, // for shift
600 x86.REG_DX, // for divide, context
601 x86.REG_SP, // for stack
608 func gconreg(as obj.As, c int64, reg int) {
612 gc.Nodconst(&n1, gc.Types[gc.TINT64], c)
613 gc.Nodreg(&n2, gc.Types[gc.TINT64], reg)
621 func ginscon(as obj.As, c int64, n2 *gc.Node) {
623 gc.Nodconst(&n1, gc.Types[gc.TINT32], c)
627 func ginscmp(op gc.Op, t *gc.Type, n1, n2 *gc.Node, likely int) *obj.Prog {
628 if t.IsInteger() || t.Etype == gc.Tptr {
629 if (n1.Op == gc.OLITERAL || n1.Op == gc.OADDR && n1.Left.Op == gc.ONAME) && n2.Op != gc.OLITERAL {
630 // Reverse comparison to place constant (including address constant) last.
637 var r1, r2, g1, g2 gc.Node
639 // A special case to make write barriers more efficient.
640 // Comparing the first field of a named struct can be done directly.
642 if n1.Op == gc.ODOT && n1.Left.Type.IsStruct() && n1.Left.Type.Field(0).Sym == n1.Sym {
646 if base.Op == gc.ONAME && base.Class&gc.PHEAP == 0 || n1.Op == gc.OINDREG {
649 gc.Regalloc(&r1, t, n1)
650 gc.Regalloc(&g1, n1.Type, &r1)
654 if n2.Op == gc.OLITERAL && t.IsInteger() || n2.Op == gc.OADDR && n2.Left.Op == gc.ONAME && n2.Left.Class == gc.PEXTERN {
657 gc.Regalloc(&r2, t, n2)
658 gc.Regalloc(&g2, n1.Type, &r2)
662 gins(optoas(gc.OCMP, t), &r1, &r2)
663 if r1.Op == gc.OREGISTER {
667 if r2.Op == gc.OREGISTER {
671 return gc.Gbranch(optoas(op, t), nil, likely)
677 func nswap(a *gc.Node, b *gc.Node) {
684 * return constant i node.
685 * overwritten by next call, but useful in calls to gins.
690 func ncon(i uint32) *gc.Node {
691 if ncon_n.Type == nil {
692 gc.Nodconst(&ncon_n, gc.Types[gc.TUINT32], 0)
694 ncon_n.SetInt(int64(i))
698 var sclean [10]gc.Node
703 * n is a 64-bit value. fill in lo and hi to refer to its 32-bit halves.
705 func split64(n *gc.Node, lo *gc.Node, hi *gc.Node) {
706 if !gc.Is64(n.Type) {
707 gc.Fatalf("split64 %v", n.Type)
710 if nsclean >= len(sclean) {
711 gc.Fatalf("split64 clean")
713 sclean[nsclean].Op = gc.OEMPTY
720 if !dotaddable(n, &n1) {
722 sclean[nsclean-1] = n1
728 if n.Class == gc.PPARAMREF {
730 gc.Cgen(n.Name.Heapaddr, &n1)
731 sclean[nsclean-1] = n1
742 lo.Type = gc.Types[gc.TUINT32]
743 if n.Type.Etype == gc.TINT64 {
744 hi.Type = gc.Types[gc.TINT32]
746 hi.Type = gc.Types[gc.TUINT32]
752 n.Convconst(&n1, n.Type)
754 gc.Nodconst(lo, gc.Types[gc.TUINT32], int64(uint32(i)))
756 if n.Type.Etype == gc.TINT64 {
757 gc.Nodconst(hi, gc.Types[gc.TINT32], int64(int32(i)))
759 gc.Nodconst(hi, gc.Types[gc.TUINT32], int64(uint32(i)))
766 gc.Fatalf("splitclean")
769 if sclean[nsclean].Op != gc.OEMPTY {
770 gc.Regfree(&sclean[nsclean])
774 // set up nodes representing fp constants
788 gc.Nodconst(&zerof, gc.Types[gc.TINT64], 0)
789 zerof.Convconst(&zerof, gc.Types[gc.TFLOAT64])
796 gc.Nodconst(&bigi, gc.Types[gc.TUINT64], 0)
798 bigi.Convconst(&two63f, gc.Types[gc.TFLOAT64])
800 gc.Nodconst(&bigi, gc.Types[gc.TUINT64], 0)
803 bigi.Convconst(&two64f, gc.Types[gc.TFLOAT64])
806 func memname(n *gc.Node, t *gc.Type) {
808 n.Sym = gc.Lookup("." + n.Sym.Name[1:]) // keep optimizer from registerizing
812 func gmove(f *gc.Node, t *gc.Node) {
813 if gc.Debug['M'] != 0 {
814 fmt.Printf("gmove %v -> %v\n", f, t)
817 ft := gc.Simsimtype(f.Type)
818 tt := gc.Simsimtype(t.Type)
821 if gc.Iscomplex[ft] || gc.Iscomplex[tt] {
826 if gc.Isfloat[ft] || gc.Isfloat[tt] {
831 // cannot have two integer memory operands;
832 // except 64-bit, which always copies via registers anyway.
835 if gc.Isint[ft] && gc.Isint[tt] && !gc.Is64(f.Type) && !gc.Is64(t.Type) && gc.Ismem(f) && gc.Ismem(t) {
839 // convert constant to desired type
840 if f.Op == gc.OLITERAL {
842 f.Convconst(&con, t.Type)
844 ft = gc.Simsimtype(con.Type)
847 // value -> value copy, only one memory operand.
848 // figure out the instruction to use.
849 // break out of switch for one-instruction gins.
850 // goto rdst for "destination must be register".
851 // goto hard for "convert to cvt type first".
852 // otherwise handle and return.
854 switch uint32(ft)<<16 | uint32(tt) {
857 gc.Fatalf("gmove %v -> %v", f, t)
861 * integer copy and truncate
863 case gc.TINT8<<16 | gc.TINT8, // same size
864 gc.TINT8<<16 | gc.TUINT8,
865 gc.TUINT8<<16 | gc.TINT8,
866 gc.TUINT8<<16 | gc.TUINT8:
869 case gc.TINT16<<16 | gc.TINT8, // truncate
870 gc.TUINT16<<16 | gc.TINT8,
871 gc.TINT32<<16 | gc.TINT8,
872 gc.TUINT32<<16 | gc.TINT8,
873 gc.TINT16<<16 | gc.TUINT8,
874 gc.TUINT16<<16 | gc.TUINT8,
875 gc.TINT32<<16 | gc.TUINT8,
876 gc.TUINT32<<16 | gc.TUINT8:
881 case gc.TINT64<<16 | gc.TINT8, // truncate low word
882 gc.TUINT64<<16 | gc.TINT8,
883 gc.TINT64<<16 | gc.TUINT8,
884 gc.TUINT64<<16 | gc.TUINT8:
887 split64(f, &flo, &fhi)
890 gc.Nodreg(&r1, t.Type, x86.REG_AX)
892 gins(x86.AMOVB, &r1, t)
896 case gc.TINT16<<16 | gc.TINT16, // same size
897 gc.TINT16<<16 | gc.TUINT16,
898 gc.TUINT16<<16 | gc.TINT16,
899 gc.TUINT16<<16 | gc.TUINT16:
902 case gc.TINT32<<16 | gc.TINT16, // truncate
903 gc.TUINT32<<16 | gc.TINT16,
904 gc.TINT32<<16 | gc.TUINT16,
905 gc.TUINT32<<16 | gc.TUINT16:
910 case gc.TINT64<<16 | gc.TINT16, // truncate low word
911 gc.TUINT64<<16 | gc.TINT16,
912 gc.TINT64<<16 | gc.TUINT16,
913 gc.TUINT64<<16 | gc.TUINT16:
916 split64(f, &flo, &fhi)
919 gc.Nodreg(&r1, t.Type, x86.REG_AX)
921 gins(x86.AMOVW, &r1, t)
925 case gc.TINT32<<16 | gc.TINT32, // same size
926 gc.TINT32<<16 | gc.TUINT32,
927 gc.TUINT32<<16 | gc.TINT32,
928 gc.TUINT32<<16 | gc.TUINT32:
931 case gc.TINT64<<16 | gc.TINT32, // truncate
932 gc.TUINT64<<16 | gc.TINT32,
933 gc.TINT64<<16 | gc.TUINT32,
934 gc.TUINT64<<16 | gc.TUINT32:
937 split64(f, &flo, &fhi)
940 gc.Nodreg(&r1, t.Type, x86.REG_AX)
942 gins(x86.AMOVL, &r1, t)
946 case gc.TINT64<<16 | gc.TINT64, // same size
947 gc.TINT64<<16 | gc.TUINT64,
948 gc.TUINT64<<16 | gc.TINT64,
949 gc.TUINT64<<16 | gc.TUINT64:
952 split64(f, &flo, &fhi)
956 split64(t, &tlo, &thi)
957 if f.Op == gc.OLITERAL {
958 gins(x86.AMOVL, &flo, &tlo)
959 gins(x86.AMOVL, &fhi, &thi)
961 // Implementation of conversion-free x = y for int64 or uint64 x.
962 // This is generated by the code that copies small values out of closures,
963 // and that code has DX live, so avoid DX and just use AX twice.
965 gc.Nodreg(&r1, gc.Types[gc.TUINT32], x86.REG_AX)
966 gins(x86.AMOVL, &flo, &r1)
967 gins(x86.AMOVL, &r1, &tlo)
968 gins(x86.AMOVL, &fhi, &r1)
969 gins(x86.AMOVL, &r1, &thi)
977 * integer up-conversions
979 case gc.TINT8<<16 | gc.TINT16, // sign extend int8
980 gc.TINT8<<16 | gc.TUINT16:
985 case gc.TINT8<<16 | gc.TINT32,
986 gc.TINT8<<16 | gc.TUINT32:
990 case gc.TINT8<<16 | gc.TINT64, // convert via int32
991 gc.TINT8<<16 | gc.TUINT64:
992 cvt = gc.Types[gc.TINT32]
996 case gc.TUINT8<<16 | gc.TINT16, // zero extend uint8
997 gc.TUINT8<<16 | gc.TUINT16:
1002 case gc.TUINT8<<16 | gc.TINT32,
1003 gc.TUINT8<<16 | gc.TUINT32:
1007 case gc.TUINT8<<16 | gc.TINT64, // convert via uint32
1008 gc.TUINT8<<16 | gc.TUINT64:
1009 cvt = gc.Types[gc.TUINT32]
1013 case gc.TINT16<<16 | gc.TINT32, // sign extend int16
1014 gc.TINT16<<16 | gc.TUINT32:
1019 case gc.TINT16<<16 | gc.TINT64, // convert via int32
1020 gc.TINT16<<16 | gc.TUINT64:
1021 cvt = gc.Types[gc.TINT32]
1025 case gc.TUINT16<<16 | gc.TINT32, // zero extend uint16
1026 gc.TUINT16<<16 | gc.TUINT32:
1031 case gc.TUINT16<<16 | gc.TINT64, // convert via uint32
1032 gc.TUINT16<<16 | gc.TUINT64:
1033 cvt = gc.Types[gc.TUINT32]
1037 case gc.TINT32<<16 | gc.TINT64, // sign extend int32
1038 gc.TINT32<<16 | gc.TUINT64:
1041 split64(t, &tlo, &thi)
1044 gc.Nodreg(&flo, tlo.Type, x86.REG_AX)
1046 gc.Nodreg(&fhi, thi.Type, x86.REG_DX)
1048 gins(x86.ACDQ, nil, nil)
1049 gins(x86.AMOVL, &flo, &tlo)
1050 gins(x86.AMOVL, &fhi, &thi)
1054 case gc.TUINT32<<16 | gc.TINT64, // zero extend uint32
1055 gc.TUINT32<<16 | gc.TUINT64:
1058 split64(t, &tlo, &thi)
1061 gins(x86.AMOVL, ncon(0), &thi)
1069 // requires register source
1071 gc.Regalloc(&r1, f.Type, t)
1078 // requires register destination
1081 gc.Regalloc(&r1, t.Type, t)
1089 // requires register intermediate
1091 gc.Regalloc(&r1, cvt, t)
1099 func floatmove(f *gc.Node, t *gc.Node) {
1102 ft := gc.Simsimtype(f.Type)
1103 tt := gc.Simsimtype(t.Type)
1106 // cannot have two floating point memory operands.
1107 if gc.Isfloat[ft] && gc.Isfloat[tt] && gc.Ismem(f) && gc.Ismem(t) {
1111 // convert constant to desired type
1112 if f.Op == gc.OLITERAL {
1114 f.Convconst(&con, t.Type)
1116 ft = gc.Simsimtype(con.Type)
1118 // some constants can't move directly to memory.
1120 // float constants come from memory.
1127 // value -> value copy, only one memory operand.
1128 // figure out the instruction to use.
1129 // break out of switch for one-instruction gins.
1130 // goto rdst for "destination must be register".
1131 // goto hard for "convert to cvt type first".
1132 // otherwise handle and return.
1134 switch uint32(ft)<<16 | uint32(tt) {
1136 if gc.Thearch.Use387 {
1143 // float to very long integer.
1144 case gc.TFLOAT32<<16 | gc.TINT64,
1145 gc.TFLOAT64<<16 | gc.TINT64:
1146 if f.Op == gc.OREGISTER {
1152 gc.Nodreg(&r1, gc.Types[ft], x86.REG_F0)
1153 if ft == gc.TFLOAT32 {
1154 gins(x86.AFMOVF, f, &r1)
1156 gins(x86.AFMOVD, f, &r1)
1159 // set round to zero mode during conversion
1161 memname(&t1, gc.Types[gc.TUINT16])
1164 memname(&t2, gc.Types[gc.TUINT16])
1165 gins(x86.AFSTCW, nil, &t1)
1166 gins(x86.AMOVW, ncon(0xf7f), &t2)
1167 gins(x86.AFLDCW, &t2, nil)
1168 if tt == gc.TINT16 {
1169 gins(x86.AFMOVWP, &r1, t)
1170 } else if tt == gc.TINT32 {
1171 gins(x86.AFMOVLP, &r1, t)
1173 gins(x86.AFMOVVP, &r1, t)
1175 gins(x86.AFLDCW, &t1, nil)
1178 case gc.TFLOAT32<<16 | gc.TUINT64,
1179 gc.TFLOAT64<<16 | gc.TUINT64:
1187 gc.Nodreg(&f0, gc.Types[ft], x86.REG_F0)
1189 gc.Nodreg(&f1, gc.Types[ft], x86.REG_F0+1)
1191 gc.Nodreg(&ax, gc.Types[gc.TUINT16], x86.REG_AX)
1193 if ft == gc.TFLOAT32 {
1194 gins(x86.AFMOVF, f, &f0)
1196 gins(x86.AFMOVD, f, &f0)
1199 // if 0 > v { answer = 0 }
1200 gins(x86.AFMOVD, &zerof, &f0)
1201 gins(x86.AFUCOMP, &f0, &f1)
1202 gins(x86.AFSTSW, nil, &ax)
1203 gins(x86.ASAHF, nil, nil)
1204 p1 := gc.Gbranch(optoas(gc.OGT, gc.Types[tt]), nil, 0)
1206 // if 1<<64 <= v { answer = 0 too }
1207 gins(x86.AFMOVD, &two64f, &f0)
1209 gins(x86.AFUCOMP, &f0, &f1)
1210 gins(x86.AFSTSW, nil, &ax)
1211 gins(x86.ASAHF, nil, nil)
1212 p2 := gc.Gbranch(optoas(gc.OGT, gc.Types[tt]), nil, 0)
1214 gins(x86.AFMOVVP, &f0, t) // don't care about t, but will pop the stack
1217 split64(t, &tlo, &thi)
1218 gins(x86.AMOVL, ncon(0), &tlo)
1219 gins(x86.AMOVL, ncon(0), &thi)
1221 p1 = gc.Gbranch(obj.AJMP, nil, 0)
1224 // in range; algorithm is:
1225 // if small enough, use native float64 -> int64 conversion.
1226 // otherwise, subtract 2^63, convert, and add it back.
1228 // set round to zero mode during conversion
1230 memname(&t1, gc.Types[gc.TUINT16])
1233 memname(&t2, gc.Types[gc.TUINT16])
1234 gins(x86.AFSTCW, nil, &t1)
1235 gins(x86.AMOVW, ncon(0xf7f), &t2)
1236 gins(x86.AFLDCW, &t2, nil)
1239 gins(x86.AFMOVD, &two63f, &f0)
1241 gins(x86.AFUCOMP, &f0, &f1)
1242 gins(x86.AFSTSW, nil, &ax)
1243 gins(x86.ASAHF, nil, nil)
1244 p2 = gc.Gbranch(optoas(gc.OLE, gc.Types[tt]), nil, 0)
1245 gins(x86.AFMOVVP, &f0, t)
1246 p3 := gc.Gbranch(obj.AJMP, nil, 0)
1248 gins(x86.AFMOVD, &two63f, &f0)
1249 gins(x86.AFSUBDP, &f0, &f1)
1250 gins(x86.AFMOVVP, &f0, t)
1251 split64(t, &tlo, &thi)
1252 gins(x86.AXORL, ncon(0x80000000), &thi) // + 2^63
1256 // restore rounding mode
1257 gins(x86.AFLDCW, &t1, nil)
1265 case gc.TINT64<<16 | gc.TFLOAT32,
1266 gc.TINT64<<16 | gc.TFLOAT64:
1267 if t.Op == gc.OREGISTER {
1271 gc.Nodreg(&f0, t.Type, x86.REG_F0)
1272 gins(x86.AFMOVV, f, &f0)
1273 if tt == gc.TFLOAT32 {
1274 gins(x86.AFMOVFP, &f0, t)
1276 gins(x86.AFMOVDP, &f0, t)
1281 // if small enough, use native int64 -> float64 conversion.
1282 // otherwise, halve (rounding to odd?), convert, and double.
1283 case gc.TUINT64<<16 | gc.TFLOAT32,
1284 gc.TUINT64<<16 | gc.TFLOAT64:
1286 gc.Nodreg(&ax, gc.Types[gc.TUINT32], x86.REG_AX)
1289 gc.Nodreg(&dx, gc.Types[gc.TUINT32], x86.REG_DX)
1291 gc.Nodreg(&cx, gc.Types[gc.TUINT32], x86.REG_CX)
1293 gc.Tempname(&t1, f.Type)
1296 split64(&t1, &tlo, &thi)
1298 gins(x86.ACMPL, &thi, ncon(0))
1299 p1 := gc.Gbranch(x86.AJLT, nil, 0)
1303 gc.Nodreg(&r1, gc.Types[tt], x86.REG_F0)
1305 gins(x86.AFMOVV, &t1, &r1)
1306 if tt == gc.TFLOAT32 {
1307 gins(x86.AFMOVFP, &r1, t)
1309 gins(x86.AFMOVDP, &r1, t)
1311 p2 := gc.Gbranch(obj.AJMP, nil, 0)
1318 p1 = gins(x86.ASHRL, ncon(1), &ax)
1319 p1.From.Index = x86.REG_DX // double-width shift DX -> AX
1321 gins(x86.AMOVL, ncon(0), &cx)
1322 gins(x86.ASETCC, nil, &cx)
1323 gins(x86.AORL, &cx, &ax)
1324 gins(x86.ASHRL, ncon(1), &dx)
1327 gc.Nodreg(&r1, gc.Types[tt], x86.REG_F0)
1329 gc.Nodreg(&r2, gc.Types[tt], x86.REG_F0+1)
1330 gins(x86.AFMOVV, &t1, &r1)
1331 gins(x86.AFMOVD, &r1, &r1)
1332 gins(x86.AFADDDP, &r1, &r2)
1333 if tt == gc.TFLOAT32 {
1334 gins(x86.AFMOVFP, &r1, t)
1336 gins(x86.AFMOVDP, &r1, t)
1343 // requires register intermediate
1345 gc.Regalloc(&r1, cvt, t)
1352 // requires memory intermediate
1354 gc.Tempname(&r1, cvt)
1361 func floatmove_387(f *gc.Node, t *gc.Node) {
1365 ft := gc.Simsimtype(f.Type)
1366 tt := gc.Simsimtype(t.Type)
1369 switch uint32(ft)<<16 | uint32(tt) {
1376 case gc.TFLOAT32<<16 | gc.TINT16,
1377 gc.TFLOAT32<<16 | gc.TINT32,
1378 gc.TFLOAT32<<16 | gc.TINT64,
1379 gc.TFLOAT64<<16 | gc.TINT16,
1380 gc.TFLOAT64<<16 | gc.TINT32,
1381 gc.TFLOAT64<<16 | gc.TINT64:
1382 if t.Op == gc.OREGISTER {
1386 gc.Nodreg(&r1, gc.Types[ft], x86.REG_F0)
1387 if f.Op != gc.OREGISTER {
1388 if ft == gc.TFLOAT32 {
1389 gins(x86.AFMOVF, f, &r1)
1391 gins(x86.AFMOVD, f, &r1)
1395 // set round to zero mode during conversion
1397 memname(&t1, gc.Types[gc.TUINT16])
1400 memname(&t2, gc.Types[gc.TUINT16])
1401 gins(x86.AFSTCW, nil, &t1)
1402 gins(x86.AMOVW, ncon(0xf7f), &t2)
1403 gins(x86.AFLDCW, &t2, nil)
1404 if tt == gc.TINT16 {
1405 gins(x86.AFMOVWP, &r1, t)
1406 } else if tt == gc.TINT32 {
1407 gins(x86.AFMOVLP, &r1, t)
1409 gins(x86.AFMOVVP, &r1, t)
1411 gins(x86.AFLDCW, &t1, nil)
1414 // convert via int32.
1415 case gc.TFLOAT32<<16 | gc.TINT8,
1416 gc.TFLOAT32<<16 | gc.TUINT16,
1417 gc.TFLOAT32<<16 | gc.TUINT8,
1418 gc.TFLOAT64<<16 | gc.TINT8,
1419 gc.TFLOAT64<<16 | gc.TUINT16,
1420 gc.TFLOAT64<<16 | gc.TUINT8:
1422 gc.Tempname(&t1, gc.Types[gc.TINT32])
1427 gc.Fatalf("gmove %v", t)
1430 gins(x86.ACMPL, &t1, ncon(-0x80&(1<<32-1)))
1431 p1 := gc.Gbranch(optoas(gc.OLT, gc.Types[gc.TINT32]), nil, -1)
1432 gins(x86.ACMPL, &t1, ncon(0x7f))
1433 p2 := gc.Gbranch(optoas(gc.OGT, gc.Types[gc.TINT32]), nil, -1)
1434 p3 := gc.Gbranch(obj.AJMP, nil, 0)
1437 gmove(ncon(-0x80&(1<<32-1)), &t1)
1442 gins(x86.ATESTL, ncon(0xffffff00), &t1)
1443 p1 := gc.Gbranch(x86.AJEQ, nil, +1)
1444 gins(x86.AMOVL, ncon(0), &t1)
1449 gins(x86.ATESTL, ncon(0xffff0000), &t1)
1450 p1 := gc.Gbranch(x86.AJEQ, nil, +1)
1451 gins(x86.AMOVL, ncon(0), &t1)
1458 // convert via int64.
1459 case gc.TFLOAT32<<16 | gc.TUINT32,
1460 gc.TFLOAT64<<16 | gc.TUINT32:
1461 cvt = gc.Types[gc.TINT64]
1468 case gc.TINT16<<16 | gc.TFLOAT32,
1469 gc.TINT16<<16 | gc.TFLOAT64,
1470 gc.TINT32<<16 | gc.TFLOAT32,
1471 gc.TINT32<<16 | gc.TFLOAT64,
1472 gc.TINT64<<16 | gc.TFLOAT32,
1473 gc.TINT64<<16 | gc.TFLOAT64:
1474 if t.Op != gc.OREGISTER {
1477 if f.Op == gc.OREGISTER {
1493 // convert via int32 memory
1494 case gc.TINT8<<16 | gc.TFLOAT32,
1495 gc.TINT8<<16 | gc.TFLOAT64,
1496 gc.TUINT16<<16 | gc.TFLOAT32,
1497 gc.TUINT16<<16 | gc.TFLOAT64,
1498 gc.TUINT8<<16 | gc.TFLOAT32,
1499 gc.TUINT8<<16 | gc.TFLOAT64:
1500 cvt = gc.Types[gc.TINT32]
1504 // convert via int64 memory
1505 case gc.TUINT32<<16 | gc.TFLOAT32,
1506 gc.TUINT32<<16 | gc.TFLOAT64:
1507 cvt = gc.Types[gc.TINT64]
1511 // The way the code generator uses floating-point
1512 // registers, a move from F0 to F0 is intended as a no-op.
1513 // On the x86, it's not: it pushes a second copy of F0
1514 // on the floating point stack. So toss it away here.
1515 // Also, F0 is the *only* register we ever evaluate
1516 // into, so we should only see register/register as F0/F0.
1520 case gc.TFLOAT32<<16 | gc.TFLOAT32,
1521 gc.TFLOAT64<<16 | gc.TFLOAT64:
1522 if gc.Ismem(f) && gc.Ismem(t) {
1525 if f.Op == gc.OREGISTER && t.Op == gc.OREGISTER {
1526 if f.Reg != x86.REG_F0 || t.Reg != x86.REG_F0 {
1533 if ft == gc.TFLOAT64 {
1537 if f.Op != gc.OREGISTER || f.Reg != x86.REG_F0 {
1538 gc.Fatalf("gmove %v", f)
1541 if ft == gc.TFLOAT64 {
1546 case gc.TFLOAT32<<16 | gc.TFLOAT64:
1547 if gc.Ismem(f) && gc.Ismem(t) {
1550 if f.Op == gc.OREGISTER && t.Op == gc.OREGISTER {
1551 if f.Reg != x86.REG_F0 || t.Reg != x86.REG_F0 {
1557 if f.Op == gc.OREGISTER {
1558 gins(x86.AFMOVDP, f, t)
1560 gins(x86.AFMOVF, f, t)
1564 case gc.TFLOAT64<<16 | gc.TFLOAT32:
1565 if gc.Ismem(f) && gc.Ismem(t) {
1568 if f.Op == gc.OREGISTER && t.Op == gc.OREGISTER {
1570 gc.Tempname(&r1, gc.Types[gc.TFLOAT32])
1571 gins(x86.AFMOVFP, f, &r1)
1572 gins(x86.AFMOVF, &r1, t)
1576 if f.Op == gc.OREGISTER {
1577 gins(x86.AFMOVFP, f, t)
1579 gins(x86.AFMOVD, f, t)
1587 // requires register intermediate
1589 gc.Regalloc(&r1, cvt, t)
1596 // requires memory intermediate
1598 gc.Tempname(&r1, cvt)
1604 // should not happen
1606 gc.Fatalf("gmove %v -> %v", gc.Nconv(f, gc.FmtLong), gc.Nconv(t, gc.FmtLong))
1611 func floatmove_sse(f *gc.Node, t *gc.Node) {
1616 ft := gc.Simsimtype(f.Type)
1617 tt := gc.Simsimtype(t.Type)
1619 switch uint32(ft)<<16 | uint32(tt) {
1620 // should not happen
1622 gc.Fatalf("gmove %v -> %v", f, t)
1626 // convert via int32.
1630 case gc.TFLOAT32<<16 | gc.TINT16,
1631 gc.TFLOAT32<<16 | gc.TINT8,
1632 gc.TFLOAT32<<16 | gc.TUINT16,
1633 gc.TFLOAT32<<16 | gc.TUINT8,
1634 gc.TFLOAT64<<16 | gc.TINT16,
1635 gc.TFLOAT64<<16 | gc.TINT8,
1636 gc.TFLOAT64<<16 | gc.TUINT16,
1637 gc.TFLOAT64<<16 | gc.TUINT8:
1638 cvt = gc.Types[gc.TINT32]
1642 // convert via int64.
1643 case gc.TFLOAT32<<16 | gc.TUINT32,
1644 gc.TFLOAT64<<16 | gc.TUINT32:
1645 cvt = gc.Types[gc.TINT64]
1649 case gc.TFLOAT32<<16 | gc.TINT32:
1653 case gc.TFLOAT64<<16 | gc.TINT32:
1657 // convert via int32 memory
1661 case gc.TINT8<<16 | gc.TFLOAT32,
1662 gc.TINT8<<16 | gc.TFLOAT64,
1663 gc.TINT16<<16 | gc.TFLOAT32,
1664 gc.TINT16<<16 | gc.TFLOAT64,
1665 gc.TUINT16<<16 | gc.TFLOAT32,
1666 gc.TUINT16<<16 | gc.TFLOAT64,
1667 gc.TUINT8<<16 | gc.TFLOAT32,
1668 gc.TUINT8<<16 | gc.TFLOAT64:
1669 cvt = gc.Types[gc.TINT32]
1673 // convert via int64 memory
1674 case gc.TUINT32<<16 | gc.TFLOAT32,
1675 gc.TUINT32<<16 | gc.TFLOAT64:
1676 cvt = gc.Types[gc.TINT64]
1680 case gc.TINT32<<16 | gc.TFLOAT32:
1684 case gc.TINT32<<16 | gc.TFLOAT64:
1691 case gc.TFLOAT32<<16 | gc.TFLOAT32:
1694 case gc.TFLOAT64<<16 | gc.TFLOAT64:
1697 case gc.TFLOAT32<<16 | gc.TFLOAT64:
1701 case gc.TFLOAT64<<16 | gc.TFLOAT32:
1709 // requires register intermediate
1711 gc.Regalloc(&r1, cvt, t)
1718 // requires memory intermediate
1720 gc.Tempname(&r1, cvt)
1726 // requires register destination
1728 gc.Regalloc(&r1, t.Type, t)
1736 func samaddr(f *gc.Node, t *gc.Node) bool {
1753 * generate one instruction:
1756 func gins(as obj.As, f *gc.Node, t *gc.Node) *obj.Prog {
1757 if as == x86.AFMOVF && f != nil && f.Op == gc.OREGISTER && t != nil && t.Op == gc.OREGISTER {
1758 gc.Fatalf("gins MOVF reg, reg")
1760 if as == x86.ACVTSD2SS && f != nil && f.Op == gc.OLITERAL {
1761 gc.Fatalf("gins CVTSD2SS const")
1763 if as == x86.AMOVSD && t != nil && t.Op == gc.OREGISTER && t.Reg == x86.REG_F0 {
1764 gc.Fatalf("gins MOVSD into F0")
1767 if as == x86.AMOVL && f != nil && f.Op == gc.OADDR && f.Left.Op == gc.ONAME && f.Left.Class != gc.PEXTERN && f.Left.Class != gc.PFUNC {
1768 // Turn MOVL $xxx(FP/SP) into LEAL xxx.
1769 // These should be equivalent but most of the backend
1770 // only expects to see LEAL, because that's what we had
1771 // historically generated. Various hidden assumptions are baked in by now.
1780 if f != nil && t != nil && samaddr(f, t) {
1785 if f != nil && gc.Isconst(f, gc.CTNIL) {
1786 gc.Fatalf("gins LEAL nil %v", f.Type)
1791 gc.Naddr(&p.From, f)
1794 if gc.Debug['g'] != 0 {
1795 fmt.Printf("%v\n", p)
1810 if true && w != 0 && f != nil && (p.From.Width > int64(w) || p.To.Width > int64(w)) {
1811 gc.Dump("bad width from:", f)
1812 gc.Dump("bad width to:", t)
1813 gc.Fatalf("bad width: %v (%d, %d)\n", p, p.From.Width, p.To.Width)
1816 if p.To.Type == obj.TYPE_ADDR && w > 0 {
1817 gc.Fatalf("bad use of addr: %v", p)
1825 gc.Nodreg(®, gc.Types[gc.TINT], x86.REG_AX)
1826 gins(x86.AXCHGL, ®, ®)
1829 func dotaddable(n *gc.Node, n1 *gc.Node) bool {
1830 if n.Op != gc.ODOT {
1836 o := gc.Dotoffset(n, oary[:], &nn)
1837 if nn != nil && nn.Addable && o == 1 && oary[0] >= 0 {
1840 n1.Xoffset += oary[0]
1850 func sudoaddable(as obj.As, n *gc.Node, a *obj.Addr) bool {