mirror of
				https://github.com/qemu/qemu.git
				synced 2025-10-31 12:07:31 +00:00 
			
		
		
		
	 62c9947fb7
			
		
	
	
		62c9947fb7
		
	
	
	
	
		
			
			Based on already existing QEMU implementation created a signed 256 bit by 128 bit division needed to implement the vector divide extended signed quadword instruction from PowerISA 3.1 Signed-off-by: Lucas Mateus Castro (alqotel) <lucas.araujo@eldorado.org.br> Reviewed-by: Richard Henderson <richard.henderson@linaro.org> Message-Id: <20220525134954.85056-6-lucas.araujo@eldorado.org.br> Signed-off-by: Daniel Henrique Barboza <danielhb413@gmail.com>
		
			
				
	
	
		
			449 lines
		
	
	
		
			12 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			449 lines
		
	
	
		
			12 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
| /*
 | |
|  * Utility compute operations used by translated code.
 | |
|  *
 | |
|  * Copyright (c) 2003 Fabrice Bellard
 | |
|  * Copyright (c) 2007 Aurelien Jarno
 | |
|  *
 | |
|  * Permission is hereby granted, free of charge, to any person obtaining a copy
 | |
|  * of this software and associated documentation files (the "Software"), to deal
 | |
|  * in the Software without restriction, including without limitation the rights
 | |
|  * to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
 | |
|  * copies of the Software, and to permit persons to whom the Software is
 | |
|  * furnished to do so, subject to the following conditions:
 | |
|  *
 | |
|  * The above copyright notice and this permission notice shall be included in
 | |
|  * all copies or substantial portions of the Software.
 | |
|  *
 | |
|  * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
 | |
|  * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
 | |
|  * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
 | |
|  * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
 | |
|  * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
 | |
|  * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN
 | |
|  * THE SOFTWARE.
 | |
|  */
 | |
| 
 | |
| #include "qemu/osdep.h"
 | |
| #include "qemu/host-utils.h"
 | |
| 
 | |
| #ifndef CONFIG_INT128
 | |
| /* Long integer helpers */
 | |
| static inline void mul64(uint64_t *plow, uint64_t *phigh,
 | |
|                          uint64_t a, uint64_t b)
 | |
| {
 | |
|     typedef union {
 | |
|         uint64_t ll;
 | |
|         struct {
 | |
| #if HOST_BIG_ENDIAN
 | |
|             uint32_t high, low;
 | |
| #else
 | |
|             uint32_t low, high;
 | |
| #endif
 | |
|         } l;
 | |
|     } LL;
 | |
|     LL rl, rm, rn, rh, a0, b0;
 | |
|     uint64_t c;
 | |
| 
 | |
|     a0.ll = a;
 | |
|     b0.ll = b;
 | |
| 
 | |
|     rl.ll = (uint64_t)a0.l.low * b0.l.low;
 | |
|     rm.ll = (uint64_t)a0.l.low * b0.l.high;
 | |
|     rn.ll = (uint64_t)a0.l.high * b0.l.low;
 | |
|     rh.ll = (uint64_t)a0.l.high * b0.l.high;
 | |
| 
 | |
|     c = (uint64_t)rl.l.high + rm.l.low + rn.l.low;
 | |
|     rl.l.high = c;
 | |
|     c >>= 32;
 | |
|     c = c + rm.l.high + rn.l.high + rh.l.low;
 | |
|     rh.l.low = c;
 | |
|     rh.l.high += (uint32_t)(c >> 32);
 | |
| 
 | |
|     *plow = rl.ll;
 | |
|     *phigh = rh.ll;
 | |
| }
 | |
| 
 | |
| /* Unsigned 64x64 -> 128 multiplication */
 | |
| void mulu64 (uint64_t *plow, uint64_t *phigh, uint64_t a, uint64_t b)
 | |
| {
 | |
|     mul64(plow, phigh, a, b);
 | |
| }
 | |
| 
 | |
| /* Signed 64x64 -> 128 multiplication */
 | |
| void muls64 (uint64_t *plow, uint64_t *phigh, int64_t a, int64_t b)
 | |
| {
 | |
|     uint64_t rh;
 | |
| 
 | |
|     mul64(plow, &rh, a, b);
 | |
| 
 | |
|     /* Adjust for signs.  */
 | |
|     if (b < 0) {
 | |
|         rh -= a;
 | |
|     }
 | |
|     if (a < 0) {
 | |
|         rh -= b;
 | |
|     }
 | |
|     *phigh = rh;
 | |
| }
 | |
| 
 | |
| /*
 | |
|  * Unsigned 128-by-64 division.
 | |
|  * Returns the remainder.
 | |
|  * Returns quotient via plow and phigh.
 | |
|  * Also returns the remainder via the function return value.
 | |
|  */
 | |
| uint64_t divu128(uint64_t *plow, uint64_t *phigh, uint64_t divisor)
 | |
| {
 | |
|     uint64_t dhi = *phigh;
 | |
|     uint64_t dlo = *plow;
 | |
|     uint64_t rem, dhighest;
 | |
|     int sh;
 | |
| 
 | |
|     if (divisor == 0 || dhi == 0) {
 | |
|         *plow  = dlo / divisor;
 | |
|         *phigh = 0;
 | |
|         return dlo % divisor;
 | |
|     } else {
 | |
|         sh = clz64(divisor);
 | |
| 
 | |
|         if (dhi < divisor) {
 | |
|             if (sh != 0) {
 | |
|                 /* normalize the divisor, shifting the dividend accordingly */
 | |
|                 divisor <<= sh;
 | |
|                 dhi = (dhi << sh) | (dlo >> (64 - sh));
 | |
|                 dlo <<= sh;
 | |
|             }
 | |
| 
 | |
|             *phigh = 0;
 | |
|             *plow = udiv_qrnnd(&rem, dhi, dlo, divisor);
 | |
|         } else {
 | |
|             if (sh != 0) {
 | |
|                 /* normalize the divisor, shifting the dividend accordingly */
 | |
|                 divisor <<= sh;
 | |
|                 dhighest = dhi >> (64 - sh);
 | |
|                 dhi = (dhi << sh) | (dlo >> (64 - sh));
 | |
|                 dlo <<= sh;
 | |
| 
 | |
|                 *phigh = udiv_qrnnd(&dhi, dhighest, dhi, divisor);
 | |
|             } else {
 | |
|                 /**
 | |
|                  * dhi >= divisor
 | |
|                  * Since the MSB of divisor is set (sh == 0),
 | |
|                  * (dhi - divisor) < divisor
 | |
|                  *
 | |
|                  * Thus, the high part of the quotient is 1, and we can
 | |
|                  * calculate the low part with a single call to udiv_qrnnd
 | |
|                  * after subtracting divisor from dhi
 | |
|                  */
 | |
|                 dhi -= divisor;
 | |
|                 *phigh = 1;
 | |
|             }
 | |
| 
 | |
|             *plow = udiv_qrnnd(&rem, dhi, dlo, divisor);
 | |
|         }
 | |
| 
 | |
|         /*
 | |
|          * since the dividend/divisor might have been normalized,
 | |
|          * the remainder might also have to be shifted back
 | |
|          */
 | |
|         return rem >> sh;
 | |
|     }
 | |
| }
 | |
| 
 | |
| /*
 | |
|  * Signed 128-by-64 division.
 | |
|  * Returns quotient via plow and phigh.
 | |
|  * Also returns the remainder via the function return value.
 | |
|  */
 | |
| int64_t divs128(uint64_t *plow, int64_t *phigh, int64_t divisor)
 | |
| {
 | |
|     bool neg_quotient = false, neg_remainder = false;
 | |
|     uint64_t unsig_hi = *phigh, unsig_lo = *plow;
 | |
|     uint64_t rem;
 | |
| 
 | |
|     if (*phigh < 0) {
 | |
|         neg_quotient = !neg_quotient;
 | |
|         neg_remainder = !neg_remainder;
 | |
| 
 | |
|         if (unsig_lo == 0) {
 | |
|             unsig_hi = -unsig_hi;
 | |
|         } else {
 | |
|             unsig_hi = ~unsig_hi;
 | |
|             unsig_lo = -unsig_lo;
 | |
|         }
 | |
|     }
 | |
| 
 | |
|     if (divisor < 0) {
 | |
|         neg_quotient = !neg_quotient;
 | |
| 
 | |
|         divisor = -divisor;
 | |
|     }
 | |
| 
 | |
|     rem = divu128(&unsig_lo, &unsig_hi, (uint64_t)divisor);
 | |
| 
 | |
|     if (neg_quotient) {
 | |
|         if (unsig_lo == 0) {
 | |
|             *phigh = -unsig_hi;
 | |
|             *plow = 0;
 | |
|         } else {
 | |
|             *phigh = ~unsig_hi;
 | |
|             *plow = -unsig_lo;
 | |
|         }
 | |
|     } else {
 | |
|         *phigh = unsig_hi;
 | |
|         *plow = unsig_lo;
 | |
|     }
 | |
| 
 | |
|     if (neg_remainder) {
 | |
|         return -rem;
 | |
|     } else {
 | |
|         return rem;
 | |
|     }
 | |
| }
 | |
| #endif
 | |
| 
 | |
| /**
 | |
|  * urshift - 128-bit Unsigned Right Shift.
 | |
|  * @plow: in/out - lower 64-bit integer.
 | |
|  * @phigh: in/out - higher 64-bit integer.
 | |
|  * @shift: in - bytes to shift, between 0 and 127.
 | |
|  *
 | |
|  * Result is zero-extended and stored in plow/phigh, which are
 | |
|  * input/output variables. Shift values outside the range will
 | |
|  * be mod to 128. In other words, the caller is responsible to
 | |
|  * verify/assert both the shift range and plow/phigh pointers.
 | |
|  */
 | |
| void urshift(uint64_t *plow, uint64_t *phigh, int32_t shift)
 | |
| {
 | |
|     shift &= 127;
 | |
|     if (shift == 0) {
 | |
|         return;
 | |
|     }
 | |
| 
 | |
|     uint64_t h = *phigh >> (shift & 63);
 | |
|     if (shift >= 64) {
 | |
|         *plow = h;
 | |
|         *phigh = 0;
 | |
|     } else {
 | |
|         *plow = (*plow >> (shift & 63)) | (*phigh << (64 - (shift & 63)));
 | |
|         *phigh = h;
 | |
|     }
 | |
| }
 | |
| 
 | |
| /**
 | |
|  * ulshift - 128-bit Unsigned Left Shift.
 | |
|  * @plow: in/out - lower 64-bit integer.
 | |
|  * @phigh: in/out - higher 64-bit integer.
 | |
|  * @shift: in - bytes to shift, between 0 and 127.
 | |
|  * @overflow: out - true if any 1-bit is shifted out.
 | |
|  *
 | |
|  * Result is zero-extended and stored in plow/phigh, which are
 | |
|  * input/output variables. Shift values outside the range will
 | |
|  * be mod to 128. In other words, the caller is responsible to
 | |
|  * verify/assert both the shift range and plow/phigh pointers.
 | |
|  */
 | |
| void ulshift(uint64_t *plow, uint64_t *phigh, int32_t shift, bool *overflow)
 | |
| {
 | |
|     uint64_t low = *plow;
 | |
|     uint64_t high = *phigh;
 | |
| 
 | |
|     shift &= 127;
 | |
|     if (shift == 0) {
 | |
|         return;
 | |
|     }
 | |
| 
 | |
|     /* check if any bit will be shifted out */
 | |
|     urshift(&low, &high, 128 - shift);
 | |
|     if (low | high) {
 | |
|         *overflow = true;
 | |
|     }
 | |
| 
 | |
|     if (shift >= 64) {
 | |
|         *phigh = *plow << (shift & 63);
 | |
|         *plow = 0;
 | |
|     } else {
 | |
|         *phigh = (*plow >> (64 - (shift & 63))) | (*phigh << (shift & 63));
 | |
|         *plow = *plow << shift;
 | |
|     }
 | |
| }
 | |
| 
 | |
| /*
 | |
|  * Unsigned 256-by-128 division.
 | |
|  * Returns the remainder via r.
 | |
|  * Returns lower 128 bit of quotient.
 | |
|  * Needs a normalized divisor (most significant bit set to 1).
 | |
|  *
 | |
|  * Adapted from include/qemu/host-utils.h udiv_qrnnd,
 | |
|  * from the GNU Multi Precision Library - longlong.h __udiv_qrnnd
 | |
|  * (https://gmplib.org/repo/gmp/file/tip/longlong.h)
 | |
|  *
 | |
|  * Licensed under the GPLv2/LGPLv3
 | |
|  */
 | |
| static Int128 udiv256_qrnnd(Int128 *r, Int128 n1, Int128 n0, Int128 d)
 | |
| {
 | |
|     Int128 d0, d1, q0, q1, r1, r0, m;
 | |
|     uint64_t mp0, mp1;
 | |
| 
 | |
|     d0 = int128_make64(int128_getlo(d));
 | |
|     d1 = int128_make64(int128_gethi(d));
 | |
| 
 | |
|     r1 = int128_remu(n1, d1);
 | |
|     q1 = int128_divu(n1, d1);
 | |
|     mp0 = int128_getlo(q1);
 | |
|     mp1 = int128_gethi(q1);
 | |
|     mulu128(&mp0, &mp1, int128_getlo(d0));
 | |
|     m = int128_make128(mp0, mp1);
 | |
|     r1 = int128_make128(int128_gethi(n0), int128_getlo(r1));
 | |
|     if (int128_ult(r1, m)) {
 | |
|         q1 = int128_sub(q1, int128_one());
 | |
|         r1 = int128_add(r1, d);
 | |
|         if (int128_uge(r1, d)) {
 | |
|             if (int128_ult(r1, m)) {
 | |
|                 q1 = int128_sub(q1, int128_one());
 | |
|                 r1 = int128_add(r1, d);
 | |
|             }
 | |
|         }
 | |
|     }
 | |
|     r1 = int128_sub(r1, m);
 | |
| 
 | |
|     r0 = int128_remu(r1, d1);
 | |
|     q0 = int128_divu(r1, d1);
 | |
|     mp0 = int128_getlo(q0);
 | |
|     mp1 = int128_gethi(q0);
 | |
|     mulu128(&mp0, &mp1, int128_getlo(d0));
 | |
|     m = int128_make128(mp0, mp1);
 | |
|     r0 = int128_make128(int128_getlo(n0), int128_getlo(r0));
 | |
|     if (int128_ult(r0, m)) {
 | |
|         q0 = int128_sub(q0, int128_one());
 | |
|         r0 = int128_add(r0, d);
 | |
|         if (int128_uge(r0, d)) {
 | |
|             if (int128_ult(r0, m)) {
 | |
|                 q0 = int128_sub(q0, int128_one());
 | |
|                 r0 = int128_add(r0, d);
 | |
|             }
 | |
|         }
 | |
|     }
 | |
|     r0 = int128_sub(r0, m);
 | |
| 
 | |
|     *r = r0;
 | |
|     return int128_or(int128_lshift(q1, 64), q0);
 | |
| }
 | |
| 
 | |
| /*
 | |
|  * Unsigned 256-by-128 division.
 | |
|  * Returns the remainder.
 | |
|  * Returns quotient via plow and phigh.
 | |
|  * Also returns the remainder via the function return value.
 | |
|  */
 | |
| Int128 divu256(Int128 *plow, Int128 *phigh, Int128 divisor)
 | |
| {
 | |
|     Int128 dhi = *phigh;
 | |
|     Int128 dlo = *plow;
 | |
|     Int128 rem, dhighest;
 | |
|     int sh;
 | |
| 
 | |
|     if (!int128_nz(divisor) || !int128_nz(dhi)) {
 | |
|         *plow  = int128_divu(dlo, divisor);
 | |
|         *phigh = int128_zero();
 | |
|         return int128_remu(dlo, divisor);
 | |
|     } else {
 | |
|         sh = clz128(divisor);
 | |
| 
 | |
|         if (int128_ult(dhi, divisor)) {
 | |
|             if (sh != 0) {
 | |
|                 /* normalize the divisor, shifting the dividend accordingly */
 | |
|                 divisor = int128_lshift(divisor, sh);
 | |
|                 dhi = int128_or(int128_lshift(dhi, sh),
 | |
|                                 int128_urshift(dlo, (128 - sh)));
 | |
|                 dlo = int128_lshift(dlo, sh);
 | |
|             }
 | |
| 
 | |
|             *phigh = int128_zero();
 | |
|             *plow = udiv256_qrnnd(&rem, dhi, dlo, divisor);
 | |
|         } else {
 | |
|             if (sh != 0) {
 | |
|                 /* normalize the divisor, shifting the dividend accordingly */
 | |
|                 divisor = int128_lshift(divisor, sh);
 | |
|                 dhighest = int128_rshift(dhi, (128 - sh));
 | |
|                 dhi = int128_or(int128_lshift(dhi, sh),
 | |
|                                 int128_urshift(dlo, (128 - sh)));
 | |
|                 dlo = int128_lshift(dlo, sh);
 | |
| 
 | |
|                 *phigh = udiv256_qrnnd(&dhi, dhighest, dhi, divisor);
 | |
|             } else {
 | |
|                 /*
 | |
|                  * dhi >= divisor
 | |
|                  * Since the MSB of divisor is set (sh == 0),
 | |
|                  * (dhi - divisor) < divisor
 | |
|                  *
 | |
|                  * Thus, the high part of the quotient is 1, and we can
 | |
|                  * calculate the low part with a single call to udiv_qrnnd
 | |
|                  * after subtracting divisor from dhi
 | |
|                  */
 | |
|                 dhi = int128_sub(dhi, divisor);
 | |
|                 *phigh = int128_one();
 | |
|             }
 | |
| 
 | |
|             *plow = udiv256_qrnnd(&rem, dhi, dlo, divisor);
 | |
|         }
 | |
| 
 | |
|         /*
 | |
|          * since the dividend/divisor might have been normalized,
 | |
|          * the remainder might also have to be shifted back
 | |
|          */
 | |
|         rem = int128_urshift(rem, sh);
 | |
|         return rem;
 | |
|     }
 | |
| }
 | |
| 
 | |
| /*
 | |
|  * Signed 256-by-128 division.
 | |
|  * Returns quotient via plow and phigh.
 | |
|  * Also returns the remainder via the function return value.
 | |
|  */
 | |
| Int128 divs256(Int128 *plow, Int128 *phigh, Int128 divisor)
 | |
| {
 | |
|     bool neg_quotient = false, neg_remainder = false;
 | |
|     Int128 unsig_hi = *phigh, unsig_lo = *plow;
 | |
|     Int128 rem;
 | |
| 
 | |
|     if (!int128_nonneg(*phigh)) {
 | |
|         neg_quotient = !neg_quotient;
 | |
|         neg_remainder = !neg_remainder;
 | |
| 
 | |
|         if (!int128_nz(unsig_lo)) {
 | |
|             unsig_hi = int128_neg(unsig_hi);
 | |
|         } else {
 | |
|             unsig_hi = int128_not(unsig_hi);
 | |
|             unsig_lo = int128_neg(unsig_lo);
 | |
|         }
 | |
|     }
 | |
| 
 | |
|     if (!int128_nonneg(divisor)) {
 | |
|         neg_quotient = !neg_quotient;
 | |
| 
 | |
|         divisor = int128_neg(divisor);
 | |
|     }
 | |
| 
 | |
|     rem = divu256(&unsig_lo, &unsig_hi, divisor);
 | |
| 
 | |
|     if (neg_quotient) {
 | |
|         if (!int128_nz(unsig_lo)) {
 | |
|             *phigh = int128_neg(unsig_hi);
 | |
|             *plow = int128_zero();
 | |
|         } else {
 | |
|             *phigh = int128_not(unsig_hi);
 | |
|             *plow = int128_neg(unsig_lo);
 | |
|         }
 | |
|     } else {
 | |
|         *phigh = unsig_hi;
 | |
|         *plow = unsig_lo;
 | |
|     }
 | |
| 
 | |
|     if (neg_remainder) {
 | |
|         return int128_neg(rem);
 | |
|     } else {
 | |
|         return rem;
 | |
|     }
 | |
| }
 |