|
Message-ID: <20210811160938.GB13220@brightrain.aerifal.cx> Date: Wed, 11 Aug 2021 12:09:39 -0400 From: Rich Felker <dalias@...c.org> To: Stefan Kanthak <stefan.kanthak@...go.de> Cc: Szabolcs Nagy <nsz@...t70.net>, musl@...ts.openwall.com Subject: Re: [PATCH] Properly simplified nextafter() On Wed, Aug 11, 2021 at 05:44:28PM +0200, Stefan Kanthak wrote: > Rich Felker <dalias@...c.org> wrote: > > > On Wed, Aug 11, 2021 at 12:53:37AM +0200, Stefan Kanthak wrote: > >> Szabolcs Nagy <nsz@...t70.net> wrote: > >> > >>>* Stefan Kanthak <stefan.kanthak@...go.de> [2021-08-10 08:23:46 +0200]: > >>>> <https://git.musl-libc.org/cgit/musl/plain/src/math/nextafter.c> > >>>> has quite some superfluous statements: > >>>> > >>>> 1. there's absolutely no need for 2 uint64_t holding |x| and |y|; > >>>> 2. IEEE-754 specifies -0.0 == +0.0, so (x == y) is equivalent to > >>>> (ax == 0) && (ay == 0): the latter 2 tests can be removed; > >>> > >>> you replaced 4 int cmps with 4 float cmps (among other things). > >> > >> and hinted that the result of the second pair of comparisions is > >> already known from the first pair. > >> > >>> it's target dependent if float compares are fast or not. > >> > >> It's also target dependent whether the floating-point registers > >> can be accessed by integer instructions, or need to be copied: > >> some win, some loose! > >> Just let the compiler/optimizer do its job! > > > > The values have been copied already to perform isnan, > > NOT necessary: the compiler may have inlined isnan() and perform > the test for example using FXAM, FUCOM or FUCOMI on i386, or > UCOMISD on AMD64, without copying the arguments. static __inline unsigned __FLOAT_BITS(float __f) { union {float __f; unsigned __i;} __u; __u.__f = __f; return __u.__i; } #define isnan(x) ( \ sizeof(x) == sizeof(float) ? (__FLOAT_BITS(x) & 0x7fffffff) > 0x7f800000 : \ sizeof(x) == sizeof(double) ? (__DOUBLE_BITS(x) & -1ULL>>1) > 0x7ffULL<<52 : \ __fpclassifyl(x) == FP_NAN) So, nope. Unless it's doing some extremely high level rewriting of this inspection of the representation. > >> 0. Doesn't musl provide target specific routines for targets with > >> soft FP? > > > > No, quite the opposite. Targets with hard fp and native insns for > > particular ops have target-specific versions, > > That's why I assumed that this may also be the case for soft FP. I don't see how that follows. Targets with soft fp are exactly the opposite of having a specialized native way to perform the operation; they do it in the most general way. > >> The code is of course smaller ... but not as small and fast as a > >> proper i386 or AMD64 assembly implementation ... which I can > >> post upon request. > > > > Full asm functions are not wanted; it's something we're trying to get > > rid of in favor of just using very small/single-insn asm statements > > with proper constraints, where it's sufficiently beneficial to have > > asm at all. But I'm not even clear how you could make this function > > more efficient with asm. The overall logic would be exactly the same > > as the C. Maybe on x86_64 there'd be some SSE instructions to let you > > elide a few things? > > No, just what the instruction set offers: 23 instructions in 72 bytes. > > nextafter: > comisd xmm1, xmm0 # CF = (from > to) > jp .Lmxcsr # from or to INDEFINITE? > je .Lequal # from = to? > sbb rdx, rdx # rdx = (from > to) ? -1 : 0 > movq rcx, xmm0 # rcx = from > mov rax, rcx > add rax, rax # CF = (from & -0.0) > jz .Lzero # from = ±0.0? > ..Lstep: > sbb rax, rax # rax = (from < 0.0) ? -1 : 0 > xor rax, rdx # rax = (from < 0.0) ^ (from > to) ? -1 : 0 > or rax, 1 # rax = (from < 0.0) ^ (from > to) ? -1 : 1 > add rax, rcx # rax = nextafter(from, to) > movq xmm0, rax # xmm0 = nextafter(from, to) > xorpd xmm1, xmm1 > ..Lmxcsr: > addsd xmm0, xmm1 # set MXCSR flags > ret > ..Lequal: > movsd xmm0, xmm1 # xmm0 = to > ret > ..Lzero: > movmskpd eax, xmm1 # rax = (to & -0.0) ? 0b?1 : 0b?0 > or eax, 2 # rax = (to & -0.0) ? 0b11 : 0b10 > ror rax, 1 # rax = (to & -0.0) ? 0x8000000000000001 : 1 > movq xmm0, rax # xmm0 = (to & -0.0) ? -0x1.0p-1074 : 0x1.0p-1074 > ret > > GCC generates here at least 12 instructions more, also longer ones, > including 2 movabs to load 0x8000000000000000 and 0x7FFFFFFFFFFFFFFF, > so the code is more than 50% fatter, mixes integer SSE and FP SSE > instructions which incur 2 cycles penalty on many Intel CPUs, with > WAY TOO MANY not so predictable (un)conditional branches. We don't use asm to optimize out 2 cycles. If the compiler is choosing a bad way to perform these loads the compiler should be fixed. But I don't think it matters in any measurable way in real usage. Rich
Powered by blists - more mailing lists
Confused about mailing lists and their use? Read about mailing lists on Wikipedia and check out these guidelines on proper formatting of your messages.