xref: /qemu/util/host-utils.c (revision 9276a31c)
1 /*
2  * Utility compute operations used by translated code.
3  *
4  * Copyright (c) 2003 Fabrice Bellard
5  * Copyright (c) 2007 Aurelien Jarno
6  *
7  * Permission is hereby granted, free of charge, to any person obtaining a copy
8  * of this software and associated documentation files (the "Software"), to deal
9  * in the Software without restriction, including without limitation the rights
10  * to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
11  * copies of the Software, and to permit persons to whom the Software is
12  * furnished to do so, subject to the following conditions:
13  *
14  * The above copyright notice and this permission notice shall be included in
15  * all copies or substantial portions of the Software.
16  *
17  * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
18  * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
19  * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
20  * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
21  * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
22  * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN
23  * THE SOFTWARE.
24  */
25 
26 #include "qemu/osdep.h"
27 #include "qemu/host-utils.h"
28 
29 #ifndef CONFIG_INT128
30 /* Long integer helpers */
31 static inline void mul64(uint64_t *plow, uint64_t *phigh,
32                          uint64_t a, uint64_t b)
33 {
34     typedef union {
35         uint64_t ll;
36         struct {
37 #ifdef HOST_WORDS_BIGENDIAN
38             uint32_t high, low;
39 #else
40             uint32_t low, high;
41 #endif
42         } l;
43     } LL;
44     LL rl, rm, rn, rh, a0, b0;
45     uint64_t c;
46 
47     a0.ll = a;
48     b0.ll = b;
49 
50     rl.ll = (uint64_t)a0.l.low * b0.l.low;
51     rm.ll = (uint64_t)a0.l.low * b0.l.high;
52     rn.ll = (uint64_t)a0.l.high * b0.l.low;
53     rh.ll = (uint64_t)a0.l.high * b0.l.high;
54 
55     c = (uint64_t)rl.l.high + rm.l.low + rn.l.low;
56     rl.l.high = c;
57     c >>= 32;
58     c = c + rm.l.high + rn.l.high + rh.l.low;
59     rh.l.low = c;
60     rh.l.high += (uint32_t)(c >> 32);
61 
62     *plow = rl.ll;
63     *phigh = rh.ll;
64 }
65 
66 /* Unsigned 64x64 -> 128 multiplication */
67 void mulu64 (uint64_t *plow, uint64_t *phigh, uint64_t a, uint64_t b)
68 {
69     mul64(plow, phigh, a, b);
70 }
71 
72 /* Signed 64x64 -> 128 multiplication */
73 void muls64 (uint64_t *plow, uint64_t *phigh, int64_t a, int64_t b)
74 {
75     uint64_t rh;
76 
77     mul64(plow, &rh, a, b);
78 
79     /* Adjust for signs.  */
80     if (b < 0) {
81         rh -= a;
82     }
83     if (a < 0) {
84         rh -= b;
85     }
86     *phigh = rh;
87 }
88 
89 /*
90  * Unsigned 128-by-64 division. Returns quotient via plow and
91  * remainder via phigh.
92  * The result must fit in 64 bits (plow) - otherwise, the result
93  * is undefined.
94  * This function will cause a division by zero if passed a zero divisor.
95  */
96 void divu128(uint64_t *plow, uint64_t *phigh, uint64_t divisor)
97 {
98     uint64_t dhi = *phigh;
99     uint64_t dlo = *plow;
100     unsigned i;
101     uint64_t carry = 0;
102 
103     if (divisor == 0 || dhi == 0) {
104         *plow  = dlo / divisor;
105         *phigh = dlo % divisor;
106     } else {
107 
108         for (i = 0; i < 64; i++) {
109             carry = dhi >> 63;
110             dhi = (dhi << 1) | (dlo >> 63);
111             if (carry || (dhi >= divisor)) {
112                 dhi -= divisor;
113                 carry = 1;
114             } else {
115                 carry = 0;
116             }
117             dlo = (dlo << 1) | carry;
118         }
119 
120         *plow = dlo;
121         *phigh = dhi;
122     }
123 }
124 
125 /*
126  * Signed 128-by-64 division. Returns quotient via plow and
127  * remainder via phigh.
128  * The result must fit in 64 bits (plow) - otherwise, the result
129  * is undefined.
130  * This function will cause a division by zero if passed a zero divisor.
131  */
132 void divs128(int64_t *plow, int64_t *phigh, int64_t divisor)
133 {
134     int sgn_dvdnd = *phigh < 0;
135     int sgn_divsr = divisor < 0;
136 
137     if (sgn_dvdnd) {
138         *plow = ~(*plow);
139         *phigh = ~(*phigh);
140         if (*plow == (int64_t)-1) {
141             *plow = 0;
142             (*phigh)++;
143          } else {
144             (*plow)++;
145          }
146     }
147 
148     if (sgn_divsr) {
149         divisor = 0 - divisor;
150     }
151 
152     divu128((uint64_t *)plow, (uint64_t *)phigh, (uint64_t)divisor);
153 
154     if (sgn_dvdnd  ^ sgn_divsr) {
155         *plow = 0 - *plow;
156     }
157 }
158 #endif
159 
160 /**
161  * urshift - 128-bit Unsigned Right Shift.
162  * @plow: in/out - lower 64-bit integer.
163  * @phigh: in/out - higher 64-bit integer.
164  * @shift: in - bytes to shift, between 0 and 127.
165  *
166  * Result is zero-extended and stored in plow/phigh, which are
167  * input/output variables. Shift values outside the range will
168  * be mod to 128. In other words, the caller is responsible to
169  * verify/assert both the shift range and plow/phigh pointers.
170  */
171 void urshift(uint64_t *plow, uint64_t *phigh, int32_t shift)
172 {
173     shift &= 127;
174     if (shift == 0) {
175         return;
176     }
177 
178     uint64_t h = *phigh >> (shift & 63);
179     if (shift >= 64) {
180         *plow = h;
181         *phigh = 0;
182     } else {
183         *plow = (*plow >> (shift & 63)) | (*phigh << (64 - (shift & 63)));
184         *phigh = h;
185     }
186 }
187 
188 /**
189  * ulshift - 128-bit Unsigned Left Shift.
190  * @plow: in/out - lower 64-bit integer.
191  * @phigh: in/out - higher 64-bit integer.
192  * @shift: in - bytes to shift, between 0 and 127.
193  * @overflow: out - true if any 1-bit is shifted out.
194  *
195  * Result is zero-extended and stored in plow/phigh, which are
196  * input/output variables. Shift values outside the range will
197  * be mod to 128. In other words, the caller is responsible to
198  * verify/assert both the shift range and plow/phigh pointers.
199  */
200 void ulshift(uint64_t *plow, uint64_t *phigh, int32_t shift, bool *overflow)
201 {
202     uint64_t low = *plow;
203     uint64_t high = *phigh;
204 
205     shift &= 127;
206     if (shift == 0) {
207         return;
208     }
209 
210     /* check if any bit will be shifted out */
211     urshift(&low, &high, 128 - shift);
212     if (low | high) {
213         *overflow = true;
214     }
215 
216     if (shift >= 64) {
217         *phigh = *plow << (shift & 63);
218         *plow = 0;
219     } else {
220         *phigh = (*plow >> (64 - (shift & 63))) | (*phigh << (shift & 63));
221         *plow = *plow << shift;
222     }
223 }
224