summaryrefslogtreecommitdiff
path: root/mpc/src/mul.c
diff options
context:
space:
mode:
Diffstat (limited to 'mpc/src/mul.c')
-rw-r--r--mpc/src/mul.c639
1 files changed, 639 insertions, 0 deletions
diff --git a/mpc/src/mul.c b/mpc/src/mul.c
new file mode 100644
index 0000000000..2be9b8d646
--- /dev/null
+++ b/mpc/src/mul.c
@@ -0,0 +1,639 @@
+/* mpc_mul -- Multiply two complex numbers
+
+Copyright (C) 2002, 2004, 2005, 2008, 2009, 2010, 2011, 2012 INRIA
+
+This file is part of GNU MPC.
+
+GNU MPC is free software; you can redistribute it and/or modify it under
+the terms of the GNU Lesser General Public License as published by the
+Free Software Foundation; either version 3 of the License, or (at your
+option) any later version.
+
+GNU MPC is distributed in the hope that it will be useful, but WITHOUT ANY
+WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS
+FOR A PARTICULAR PURPOSE. See the GNU Lesser General Public License for
+more details.
+
+You should have received a copy of the GNU Lesser General Public License
+along with this program. If not, see http://www.gnu.org/licenses/ .
+*/
+
+#include <stdio.h> /* for MPC_ASSERT */
+#include "mpc-impl.h"
+
+#define mpz_add_si(z,x,y) do { \
+ if (y >= 0) \
+ mpz_add_ui (z, x, (long int) y); \
+ else \
+ mpz_sub_ui (z, x, (long int) (-y)); \
+ } while (0);
+
+/* compute z=x*y when x has an infinite part */
+static int
+mul_infinite (mpc_ptr z, mpc_srcptr x, mpc_srcptr y)
+{
+ /* Let x=xr+i*xi and y=yr+i*yi; extract the signs of the operands */
+ int xrs = mpfr_signbit (mpc_realref (x)) ? -1 : 1;
+ int xis = mpfr_signbit (mpc_imagref (x)) ? -1 : 1;
+ int yrs = mpfr_signbit (mpc_realref (y)) ? -1 : 1;
+ int yis = mpfr_signbit (mpc_imagref (y)) ? -1 : 1;
+
+ int u, v;
+
+ /* compute the sign of
+ u = xrs * yrs * xr * yr - xis * yis * xi * yi
+ v = xrs * yis * xr * yi + xis * yrs * xi * yr
+ +1 if positive, -1 if negatiye, 0 if NaN */
+ if ( mpfr_nan_p (mpc_realref (x)) || mpfr_nan_p (mpc_imagref (x))
+ || mpfr_nan_p (mpc_realref (y)) || mpfr_nan_p (mpc_imagref (y))) {
+ u = 0;
+ v = 0;
+ }
+ else if (mpfr_inf_p (mpc_realref (x))) {
+ /* x = (+/-inf) xr + i*xi */
+ u = ( mpfr_zero_p (mpc_realref (y))
+ || (mpfr_inf_p (mpc_imagref (x)) && mpfr_zero_p (mpc_imagref (y)))
+ || (mpfr_zero_p (mpc_imagref (x)) && mpfr_inf_p (mpc_imagref (y)))
+ || ( (mpfr_inf_p (mpc_imagref (x)) || mpfr_inf_p (mpc_imagref (y)))
+ && xrs*yrs == xis*yis)
+ ? 0 : xrs * yrs);
+ v = ( mpfr_zero_p (mpc_imagref (y))
+ || (mpfr_inf_p (mpc_imagref (x)) && mpfr_zero_p (mpc_realref (y)))
+ || (mpfr_zero_p (mpc_imagref (x)) && mpfr_inf_p (mpc_realref (y)))
+ || ( (mpfr_inf_p (mpc_imagref (x)) || mpfr_inf_p (mpc_imagref (x)))
+ && xrs*yis != xis*yrs)
+ ? 0 : xrs * yis);
+ }
+ else {
+ /* x = xr + i*(+/-inf) with |xr| != inf */
+ u = ( mpfr_zero_p (mpc_imagref (y))
+ || (mpfr_zero_p (mpc_realref (x)) && mpfr_inf_p (mpc_realref (y)))
+ || (mpfr_inf_p (mpc_realref (y)) && xrs*yrs == xis*yis)
+ ? 0 : -xis * yis);
+ v = ( mpfr_zero_p (mpc_realref (y))
+ || (mpfr_zero_p (mpc_realref (x)) && mpfr_inf_p (mpc_imagref (y)))
+ || (mpfr_inf_p (mpc_imagref (y)) && xrs*yis != xis*yrs)
+ ? 0 : xis * yrs);
+ }
+
+ if (u == 0 && v == 0) {
+ /* Naive result is NaN+i*NaN. Obtain an infinity using the algorithm
+ given in Annex G.5.1 of the ISO C99 standard */
+ int xr = (mpfr_zero_p (mpc_realref (x)) || mpfr_nan_p (mpc_realref (x)) ? 0
+ : (mpfr_inf_p (mpc_realref (x)) ? 1 : 0));
+ int xi = (mpfr_zero_p (mpc_imagref (x)) || mpfr_nan_p (mpc_imagref (x)) ? 0
+ : (mpfr_inf_p (mpc_imagref (x)) ? 1 : 0));
+ int yr = (mpfr_zero_p (mpc_realref (y)) || mpfr_nan_p (mpc_realref (y)) ? 0 : 1);
+ int yi = (mpfr_zero_p (mpc_imagref (y)) || mpfr_nan_p (mpc_imagref (y)) ? 0 : 1);
+ if (mpc_inf_p (y)) {
+ yr = mpfr_inf_p (mpc_realref (y)) ? 1 : 0;
+ yi = mpfr_inf_p (mpc_imagref (y)) ? 1 : 0;
+ }
+
+ u = xrs * xr * yrs * yr - xis * xi * yis * yi;
+ v = xrs * xr * yis * yi + xis * xi * yrs * yr;
+ }
+
+ if (u == 0)
+ mpfr_set_nan (mpc_realref (z));
+ else
+ mpfr_set_inf (mpc_realref (z), u);
+
+ if (v == 0)
+ mpfr_set_nan (mpc_imagref (z));
+ else
+ mpfr_set_inf (mpc_imagref (z), v);
+
+ return MPC_INEX (0, 0); /* exact */
+}
+
+
+/* compute z = x*y for Im(y) == 0 */
+static int
+mul_real (mpc_ptr z, mpc_srcptr x, mpc_srcptr y, mpc_rnd_t rnd)
+{
+ int xrs, xis, yrs, yis;
+ int inex;
+
+ /* save signs of operands */
+ xrs = MPFR_SIGNBIT (mpc_realref (x));
+ xis = MPFR_SIGNBIT (mpc_imagref (x));
+ yrs = MPFR_SIGNBIT (mpc_realref (y));
+ yis = MPFR_SIGNBIT (mpc_imagref (y));
+
+ inex = mpc_mul_fr (z, x, mpc_realref (y), rnd);
+ /* Signs of zeroes may be wrong. Their correction does not change the
+ inexact flag. */
+ if (mpfr_zero_p (mpc_realref (z)))
+ mpfr_setsign (mpc_realref (z), mpc_realref (z), MPC_RND_RE(rnd) == GMP_RNDD
+ || (xrs != yrs && xis == yis), GMP_RNDN);
+ if (mpfr_zero_p (mpc_imagref (z)))
+ mpfr_setsign (mpc_imagref (z), mpc_imagref (z), MPC_RND_IM (rnd) == GMP_RNDD
+ || (xrs != yis && xis != yrs), GMP_RNDN);
+
+ return inex;
+}
+
+
+/* compute z = x*y for Re(y) == 0, and Im(x) != 0 and Im(y) != 0 */
+static int
+mul_imag (mpc_ptr z, mpc_srcptr x, mpc_srcptr y, mpc_rnd_t rnd)
+{
+ int sign;
+ int inex_re, inex_im;
+ int overlap = z == x || z == y;
+ mpc_t rop;
+
+ if (overlap)
+ mpc_init3 (rop, MPC_PREC_RE (z), MPC_PREC_IM (z));
+ else
+ rop [0] = z[0];
+
+ sign = (MPFR_SIGNBIT (mpc_realref (y)) != MPFR_SIGNBIT (mpc_imagref (x)))
+ && (MPFR_SIGNBIT (mpc_imagref (y)) != MPFR_SIGNBIT (mpc_realref (x)));
+
+ inex_re = -mpfr_mul (mpc_realref (rop), mpc_imagref (x), mpc_imagref (y),
+ INV_RND (MPC_RND_RE (rnd)));
+ mpfr_neg (mpc_realref (rop), mpc_realref (rop), GMP_RNDN); /* exact */
+ inex_im = mpfr_mul (mpc_imagref (rop), mpc_realref (x), mpc_imagref (y),
+ MPC_RND_IM (rnd));
+ mpc_set (z, rop, MPC_RNDNN);
+
+ /* Sign of zeroes may be wrong (note that Re(z) cannot be zero) */
+ if (mpfr_zero_p (mpc_imagref (z)))
+ mpfr_setsign (mpc_imagref (z), mpc_imagref (z), MPC_RND_IM (rnd) == GMP_RNDD
+ || sign, GMP_RNDN);
+
+ if (overlap)
+ mpc_clear (rop);
+
+ return MPC_INEX (inex_re, inex_im);
+}
+
+
+static int
+mpfr_fmma (mpfr_ptr z, mpfr_srcptr a, mpfr_srcptr b, mpfr_srcptr c,
+ mpfr_srcptr d, int sign, mpfr_rnd_t rnd)
+{
+ /* Computes z = ab+cd if sign >= 0, or z = ab-cd if sign < 0.
+ Assumes that a, b, c, d are finite and non-zero; so any multiplication
+ of two of them yielding an infinity is an overflow, and a
+ multiplication yielding 0 is an underflow.
+ Assumes further that z is distinct from a, b, c, d. */
+
+ int inex;
+ mpfr_t u, v;
+
+ /* u=a*b, v=sign*c*d exactly */
+ mpfr_init2 (u, mpfr_get_prec (a) + mpfr_get_prec (b));
+ mpfr_init2 (v, mpfr_get_prec (c) + mpfr_get_prec (d));
+ mpfr_mul (u, a, b, GMP_RNDN);
+ mpfr_mul (v, c, d, GMP_RNDN);
+ if (sign < 0)
+ mpfr_neg (v, v, GMP_RNDN);
+
+ /* tentatively compute z as u+v; here we need z to be distinct
+ from a, b, c, d to not lose the latter */
+ inex = mpfr_add (z, u, v, rnd);
+
+ if (mpfr_inf_p (z)) {
+ /* replace by "correctly rounded overflow" */
+ mpfr_set_si (z, (mpfr_signbit (z) ? -1 : 1), GMP_RNDN);
+ inex = mpfr_mul_2ui (z, z, mpfr_get_emax (), rnd);
+ }
+ else if (mpfr_zero_p (u) && !mpfr_zero_p (v)) {
+ /* exactly u underflowed, determine inexact flag */
+ inex = (mpfr_signbit (u) ? 1 : -1);
+ }
+ else if (mpfr_zero_p (v) && !mpfr_zero_p (u)) {
+ /* exactly v underflowed, determine inexact flag */
+ inex = (mpfr_signbit (v) ? 1 : -1);
+ }
+ else if (mpfr_nan_p (z) || (mpfr_zero_p (u) && mpfr_zero_p (v))) {
+ /* In the first case, u and v are infinities with opposite signs.
+ In the second case, u and v are zeroes; their sum may be 0 or the
+ least representable number, with a sign to be determined.
+ Redo the computations with mpz_t exponents */
+ mpfr_exp_t ea, eb, ec, ed;
+ mpz_t eu, ev;
+ /* cheat to work around the const qualifiers */
+
+ /* Normalise the input by shifting and keep track of the shifts in
+ the exponents of u and v */
+ ea = mpfr_get_exp (a);
+ eb = mpfr_get_exp (b);
+ ec = mpfr_get_exp (c);
+ ed = mpfr_get_exp (d);
+
+ mpfr_set_exp ((mpfr_ptr) a, (mpfr_prec_t) 0);
+ mpfr_set_exp ((mpfr_ptr) b, (mpfr_prec_t) 0);
+ mpfr_set_exp ((mpfr_ptr) c, (mpfr_prec_t) 0);
+ mpfr_set_exp ((mpfr_ptr) d, (mpfr_prec_t) 0);
+
+ mpz_init (eu);
+ mpz_init (ev);
+ mpz_set_si (eu, (long int) ea);
+ mpz_add_si (eu, eu, (long int) eb);
+ mpz_set_si (ev, (long int) ec);
+ mpz_add_si (ev, ev, (long int) ed);
+
+ /* recompute u and v and move exponents to eu and ev */
+ mpfr_mul (u, a, b, GMP_RNDN);
+ /* exponent of u is non-positive */
+ mpz_sub_ui (eu, eu, (unsigned long int) (-mpfr_get_exp (u)));
+ mpfr_set_exp (u, (mpfr_prec_t) 0);
+ mpfr_mul (v, c, d, GMP_RNDN);
+ if (sign < 0)
+ mpfr_neg (v, v, GMP_RNDN);
+ mpz_sub_ui (ev, ev, (unsigned long int) (-mpfr_get_exp (v)));
+ mpfr_set_exp (v, (mpfr_prec_t) 0);
+
+ if (mpfr_nan_p (z)) {
+ mpfr_exp_t emax = mpfr_get_emax ();
+ int overflow;
+ /* We have a = ma * 2^ea with 1/2 <= |ma| < 1 and ea <= emax, and
+ analogously for b. So eu <= 2*emax, and eu > emax since we have
+ an overflow. The same holds for ev. Shift u and v by as much as
+ possible so that one of them has exponent emax and the
+ remaining exponents in eu and ev are the same. Then carry out
+ the addition. Shifting u and v prevents an underflow. */
+ if (mpz_cmp (eu, ev) >= 0) {
+ mpfr_set_exp (u, emax);
+ mpz_sub_ui (eu, eu, (long int) emax);
+ mpz_sub (ev, ev, eu);
+ mpfr_set_exp (v, (mpfr_exp_t) mpz_get_ui (ev));
+ /* remaining common exponent is now in eu */
+ }
+ else {
+ mpfr_set_exp (v, emax);
+ mpz_sub_ui (ev, ev, (long int) emax);
+ mpz_sub (eu, eu, ev);
+ mpfr_set_exp (u, (mpfr_exp_t) mpz_get_ui (eu));
+ mpz_set (eu, ev);
+ /* remaining common exponent is now also in eu */
+ }
+ inex = mpfr_add (z, u, v, rnd);
+ /* Result is finite since u and v have different signs. */
+ overflow = mpfr_mul_2ui (z, z, mpz_get_ui (eu), rnd);
+ if (overflow)
+ inex = overflow;
+ }
+ else {
+ int underflow;
+ /* Addition of two zeroes with same sign. We have a = ma * 2^ea
+ with 1/2 <= |ma| < 1 and ea >= emin and similarly for b.
+ So 2*emin < 2*emin+1 <= eu < emin < 0, and analogously for v. */
+ mpfr_exp_t emin = mpfr_get_emin ();
+ if (mpz_cmp (eu, ev) <= 0) {
+ mpfr_set_exp (u, emin);
+ mpz_add_ui (eu, eu, (unsigned long int) (-emin));
+ mpz_sub (ev, ev, eu);
+ mpfr_set_exp (v, (mpfr_exp_t) mpz_get_si (ev));
+ }
+ else {
+ mpfr_set_exp (v, emin);
+ mpz_add_ui (ev, ev, (unsigned long int) (-emin));
+ mpz_sub (eu, eu, ev);
+ mpfr_set_exp (u, (mpfr_exp_t) mpz_get_si (eu));
+ mpz_set (eu, ev);
+ }
+ inex = mpfr_add (z, u, v, rnd);
+ mpz_neg (eu, eu);
+ underflow = mpfr_div_2ui (z, z, mpz_get_ui (eu), rnd);
+ if (underflow)
+ inex = underflow;
+ }
+
+ mpz_clear (eu);
+ mpz_clear (ev);
+
+ mpfr_set_exp ((mpfr_ptr) a, ea);
+ mpfr_set_exp ((mpfr_ptr) b, eb);
+ mpfr_set_exp ((mpfr_ptr) c, ec);
+ mpfr_set_exp ((mpfr_ptr) d, ed);
+ /* works also when some of a, b, c, d are not all distinct */
+ }
+
+ mpfr_clear (u);
+ mpfr_clear (v);
+
+ return inex;
+}
+
+
+int
+mpc_mul_naive (mpc_ptr z, mpc_srcptr x, mpc_srcptr y, mpc_rnd_t rnd)
+{
+ /* computes z=x*y by the schoolbook method, where x and y are assumed
+ to be finite and without zero parts */
+ int overlap, inex;
+ mpc_t rop;
+
+ MPC_ASSERT ( mpfr_regular_p (mpc_realref (x)) && mpfr_regular_p (mpc_imagref (x))
+ && mpfr_regular_p (mpc_realref (y)) && mpfr_regular_p (mpc_imagref (y)));
+ overlap = (z == x) || (z == y);
+ if (overlap)
+ mpc_init3 (rop, MPC_PREC_RE (z), MPC_PREC_IM (z));
+ else
+ rop [0] = z [0];
+
+ inex = MPC_INEX (mpfr_fmma (mpc_realref (rop), mpc_realref (x), mpc_realref (y), mpc_imagref (x),
+ mpc_imagref (y), -1, MPC_RND_RE (rnd)),
+ mpfr_fmma (mpc_imagref (rop), mpc_realref (x), mpc_imagref (y), mpc_imagref (x),
+ mpc_realref (y), +1, MPC_RND_IM (rnd)));
+
+ mpc_set (z, rop, MPC_RNDNN);
+ if (overlap)
+ mpc_clear (rop);
+
+ return inex;
+}
+
+
+int
+mpc_mul_karatsuba (mpc_ptr rop, mpc_srcptr op1, mpc_srcptr op2, mpc_rnd_t rnd)
+{
+ /* computes rop=op1*op2 by a Karatsuba algorithm, where op1 and op2
+ are assumed to be finite and without zero parts */
+ mpfr_srcptr a, b, c, d;
+ int mul_i, ok, inexact, mul_a, mul_c, inex_re = 0, inex_im = 0, sign_x, sign_u;
+ mpfr_t u, v, w, x;
+ mpfr_prec_t prec, prec_re, prec_u, prec_v, prec_w;
+ mpfr_rnd_t rnd_re, rnd_u;
+ int overlap;
+ /* true if rop == op1 or rop == op2 */
+ mpc_t result;
+ /* overlap is quite difficult to handle, because we have to tentatively
+ round the variable u in the end to either the real or the imaginary
+ part of rop (it is not possible to tell now whether the real or
+ imaginary part is used). If this fails, we have to start again and
+ need the correct values of op1 and op2.
+ So we just create a new variable for the result in this case. */
+ int loop;
+ const int MAX_MUL_LOOP = 1;
+
+ overlap = (rop == op1) || (rop == op2);
+ if (overlap)
+ mpc_init3 (result, MPC_PREC_RE (rop), MPC_PREC_IM (rop));
+ else
+ result [0] = rop [0];
+
+ a = mpc_realref(op1);
+ b = mpc_imagref(op1);
+ c = mpc_realref(op2);
+ d = mpc_imagref(op2);
+
+ /* (a + i*b) * (c + i*d) = [ac - bd] + i*[ad + bc] */
+
+ mul_i = 0; /* number of multiplications by i */
+ mul_a = 1; /* implicit factor for a */
+ mul_c = 1; /* implicit factor for c */
+
+ if (mpfr_cmp_abs (a, b) < 0)
+ {
+ MPFR_SWAP (a, b);
+ mul_i ++;
+ mul_a = -1; /* consider i * (a+i*b) = -b + i*a */
+ }
+
+ if (mpfr_cmp_abs (c, d) < 0)
+ {
+ MPFR_SWAP (c, d);
+ mul_i ++;
+ mul_c = -1; /* consider -d + i*c instead of c + i*d */
+ }
+
+ /* find the precision and rounding mode for the new real part */
+ if (mul_i % 2)
+ {
+ prec_re = MPC_PREC_IM(rop);
+ rnd_re = MPC_RND_IM(rnd);
+ }
+ else /* mul_i = 0 or 2 */
+ {
+ prec_re = MPC_PREC_RE(rop);
+ rnd_re = MPC_RND_RE(rnd);
+ }
+
+ if (mul_i)
+ rnd_re = INV_RND(rnd_re);
+
+ /* now |a| >= |b| and |c| >= |d| */
+ prec = MPC_MAX_PREC(rop);
+
+ mpfr_init2 (v, prec_v = mpfr_get_prec (a) + mpfr_get_prec (d));
+ mpfr_init2 (w, prec_w = mpfr_get_prec (b) + mpfr_get_prec (c));
+ mpfr_init2 (u, 2);
+ mpfr_init2 (x, 2);
+
+ inexact = mpfr_mul (v, a, d, GMP_RNDN);
+ if (inexact) {
+ /* over- or underflow */
+ ok = 0;
+ goto clear;
+ }
+ if (mul_a == -1)
+ mpfr_neg (v, v, GMP_RNDN);
+
+ inexact = mpfr_mul (w, b, c, GMP_RNDN);
+ if (inexact) {
+ /* over- or underflow */
+ ok = 0;
+ goto clear;
+ }
+ if (mul_c == -1)
+ mpfr_neg (w, w, GMP_RNDN);
+
+ /* compute sign(v-w) */
+ sign_x = mpfr_cmp_abs (v, w);
+ if (sign_x > 0)
+ sign_x = 2 * mpfr_sgn (v) - mpfr_sgn (w);
+ else if (sign_x == 0)
+ sign_x = mpfr_sgn (v) - mpfr_sgn (w);
+ else
+ sign_x = mpfr_sgn (v) - 2 * mpfr_sgn (w);
+
+ sign_u = mul_a * mpfr_sgn (a) * mul_c * mpfr_sgn (c);
+
+ if (sign_x * sign_u < 0)
+ { /* swap inputs */
+ MPFR_SWAP (a, c);
+ MPFR_SWAP (b, d);
+ mpfr_swap (v, w);
+ { int tmp; tmp = mul_a; mul_a = mul_c; mul_c = tmp; }
+ sign_x = - sign_x;
+ }
+
+ /* now sign_x * sign_u >= 0 */
+ loop = 0;
+ do
+ {
+ loop++;
+ /* the following should give failures with prob. <= 1/prec */
+ prec += mpc_ceil_log2 (prec) + 3;
+
+ mpfr_set_prec (u, prec_u = prec);
+ mpfr_set_prec (x, prec);
+
+ /* first compute away(b +/- a) and store it in u */
+ inexact = (mul_a == -1 ?
+ ROUND_AWAY (mpfr_sub (u, b, a, MPFR_RNDA), u) :
+ ROUND_AWAY (mpfr_add (u, b, a, MPFR_RNDA), u));
+
+ /* then compute away(+/-c - d) and store it in x */
+ inexact |= (mul_c == -1 ?
+ ROUND_AWAY (mpfr_add (x, c, d, MPFR_RNDA), x) :
+ ROUND_AWAY (mpfr_sub (x, c, d, MPFR_RNDA), x));
+ if (mul_c == -1)
+ mpfr_neg (x, x, GMP_RNDN);
+
+ if (inexact == 0)
+ mpfr_prec_round (u, prec_u = 2 * prec, GMP_RNDN);
+
+ /* compute away(u*x) and store it in u */
+ inexact |= ROUND_AWAY (mpfr_mul (u, u, x, MPFR_RNDA), u);
+ /* (a+b)*(c-d) */
+
+ /* if all computations are exact up to here, it may be that
+ the real part is exact, thus we need if possible to
+ compute v - w exactly */
+ if (inexact == 0)
+ {
+ mpfr_prec_t prec_x;
+ /* v and w are different from 0, so mpfr_get_exp is safe to use */
+ prec_x = SAFE_ABS (mpfr_exp_t, mpfr_get_exp (v) - mpfr_get_exp (w))
+ + MPC_MAX (prec_v, prec_w) + 1;
+ /* +1 is necessary for a potential carry */
+ /* ensure we do not use a too large precision */
+ if (prec_x > prec_u)
+ prec_x = prec_u;
+ if (prec_x > prec)
+ mpfr_prec_round (x, prec_x, GMP_RNDN);
+ }
+
+ rnd_u = (sign_u > 0) ? GMP_RNDU : GMP_RNDD;
+ inexact |= mpfr_sub (x, v, w, rnd_u); /* ad - bc */
+
+ /* in case u=0, ensure that rnd_u rounds x away from zero */
+ if (mpfr_sgn (u) == 0)
+ rnd_u = (mpfr_sgn (x) > 0) ? GMP_RNDU : GMP_RNDD;
+ inexact |= mpfr_add (u, u, x, rnd_u); /* ac - bd */
+
+ ok = inexact == 0 ||
+ mpfr_can_round (u, prec_u - 3, rnd_u, GMP_RNDZ,
+ prec_re + (rnd_re == GMP_RNDN));
+ /* this ensures both we can round correctly and determine the correct
+ inexact flag (for rounding to nearest) */
+ }
+ while (!ok && loop <= MAX_MUL_LOOP);
+ /* after MAX_MUL_LOOP rounds, use mpc_naive instead */
+
+ if (ok) {
+ /* if inexact is zero, then u is exactly ac-bd, otherwise fix the sign
+ of the inexact flag for u, which was rounded away from ac-bd */
+ if (inexact != 0)
+ inexact = mpfr_sgn (u);
+
+ if (mul_i == 0)
+ {
+ inex_re = mpfr_set (mpc_realref(result), u, MPC_RND_RE(rnd));
+ if (inex_re == 0)
+ {
+ inex_re = inexact; /* u is rounded away from 0 */
+ inex_im = mpfr_add (mpc_imagref(result), v, w, MPC_RND_IM(rnd));
+ }
+ else
+ inex_im = mpfr_add (mpc_imagref(result), v, w, MPC_RND_IM(rnd));
+ }
+ else if (mul_i == 1) /* (x+i*y)/i = y - i*x */
+ {
+ inex_im = mpfr_neg (mpc_imagref(result), u, MPC_RND_IM(rnd));
+ if (inex_im == 0)
+ {
+ inex_im = -inexact; /* u is rounded away from 0 */
+ inex_re = mpfr_add (mpc_realref(result), v, w, MPC_RND_RE(rnd));
+ }
+ else
+ inex_re = mpfr_add (mpc_realref(result), v, w, MPC_RND_RE(rnd));
+ }
+ else /* mul_i = 2, z/i^2 = -z */
+ {
+ inex_re = mpfr_neg (mpc_realref(result), u, MPC_RND_RE(rnd));
+ if (inex_re == 0)
+ {
+ inex_re = -inexact; /* u is rounded away from 0 */
+ inex_im = -mpfr_add (mpc_imagref(result), v, w,
+ INV_RND(MPC_RND_IM(rnd)));
+ mpfr_neg (mpc_imagref(result), mpc_imagref(result), MPC_RND_IM(rnd));
+ }
+ else
+ {
+ inex_im = -mpfr_add (mpc_imagref(result), v, w,
+ INV_RND(MPC_RND_IM(rnd)));
+ mpfr_neg (mpc_imagref(result), mpc_imagref(result), MPC_RND_IM(rnd));
+ }
+ }
+
+ mpc_set (rop, result, MPC_RNDNN);
+ }
+
+clear:
+ mpfr_clear (u);
+ mpfr_clear (v);
+ mpfr_clear (w);
+ mpfr_clear (x);
+ if (overlap)
+ mpc_clear (result);
+
+ if (ok)
+ return MPC_INEX(inex_re, inex_im);
+ else
+ return mpc_mul_naive (rop, op1, op2, rnd);
+}
+
+
+int
+mpc_mul (mpc_ptr a, mpc_srcptr b, mpc_srcptr c, mpc_rnd_t rnd)
+{
+ /* Conforming to ISO C99 standard (G.5.1 multiplicative operators),
+ infinities are treated specially if both parts are NaN when computed
+ naively. */
+ if (mpc_inf_p (b))
+ return mul_infinite (a, b, c);
+ if (mpc_inf_p (c))
+ return mul_infinite (a, c, b);
+
+ /* NaN contamination of both parts in result */
+ if (mpfr_nan_p (mpc_realref (b)) || mpfr_nan_p (mpc_imagref (b))
+ || mpfr_nan_p (mpc_realref (c)) || mpfr_nan_p (mpc_imagref (c))) {
+ mpfr_set_nan (mpc_realref (a));
+ mpfr_set_nan (mpc_imagref (a));
+ return MPC_INEX (0, 0);
+ }
+
+ /* check for real multiplication */
+ if (mpfr_zero_p (mpc_imagref (b)))
+ return mul_real (a, c, b, rnd);
+ if (mpfr_zero_p (mpc_imagref (c)))
+ return mul_real (a, b, c, rnd);
+
+ /* check for purely imaginary multiplication */
+ if (mpfr_zero_p (mpc_realref (b)))
+ return mul_imag (a, c, b, rnd);
+ if (mpfr_zero_p (mpc_realref (c)))
+ return mul_imag (a, b, c, rnd);
+
+ /* If the real and imaginary part of one argument have a very different */
+ /* exponent, it is not reasonable to use Karatsuba multiplication. */
+ if ( SAFE_ABS (mpfr_exp_t,
+ mpfr_get_exp (mpc_realref (b)) - mpfr_get_exp (mpc_imagref (b)))
+ > (mpfr_exp_t) MPC_MAX_PREC (b) / 2
+ || SAFE_ABS (mpfr_exp_t,
+ mpfr_get_exp (mpc_realref (c)) - mpfr_get_exp (mpc_imagref (c)))
+ > (mpfr_exp_t) MPC_MAX_PREC (c) / 2)
+ return mpc_mul_naive (a, b, c, rnd);
+ else
+ return ((MPC_MAX_PREC(a)
+ <= (mpfr_prec_t) MUL_KARATSUBA_THRESHOLD * BITS_PER_MP_LIMB)
+ ? mpc_mul_naive : mpc_mul_karatsuba) (a, b, c, rnd);
+}