|
Message-ID: <438FA5B9D0E8497FA998E44E99C7CC8B@H270> Date: Wed, 11 Dec 2019 22:17:09 +0100 From: "Stefan Kanthak" <stefan.kanthak@...go.de> To: "Szabolcs Nagy" <nsz@...t70.net>, <musl@...ts.openwall.com> Subject: Re: [PATCH] fmax(), fmaxf(), fmaxl(), fmin(), fminf(), fminl() simplified "Szabolcs Nagy" <nsz@...t70.net> wrote: >* Stefan Kanthak <stefan.kanthak@...go.de> [2019-12-11 13:33:44 +0100]: >> "Szabolcs Nagy" <nsz@...t70.net> wrote: >> >* Stefan Kanthak <stefan.kanthak@...go.de> [2019-12-11 10:55:29 +0100]: >> > these two are not equivalent for snan input, but we dont care >> > about snan, nor the compiler by default, so the compiler can >> > optimize one to the other (although musl uses explicit int >> > arithmetics instead of __builtin_isnan so it's a bit harder). >> >> The latter behaviour was my reason to use (x != x) here: I attempt >> to replace as many function calls as possible with "normal" code, >> and also try to avoid transfers to/from FPU/SSE registers to/from >> integer registers if that does not result in faster/shorter code. > > why not just change the definition of isnan then? Because I did not want to introduce such a global change; until now my patches are just local (peephole) optimisations. > #if __GNUC__ > xxx > #define isnan(x) sizeof(x)==sizeof(float) ? __builtin_isnanf(x) : ... This is better than my proposed change, as it also avoids the side- effect of (x != x) which can raise exceptions, and gets rid of the explicit transfer to integer registers, which can hurt performance. The macros isinf(), isnormal(), isfinite(), signbit() should of course be implemented in a similar way too, and the (internal only?) functions __FLOAT_BITS() and __DOUBLE_BITS() removed completely! regards Stefan PS: the following is just a "Gedankenspiel", extending the idea to avoid transfers from/to SSE registers. On x86-64, functions like isunordered(), copysign() etc. may be implemented using SSE intrinsics _mm_*() as follows: #include <immintrin.h> int signbit(double argument) { return /* 1 & */ _mm_movemask_pd(_mm_set_sd(argument)); } int isunordered(double a, double b) { #if 0 return _mm_comieq_sd(_mm_cmp_sd(_mm_set_sd(a), _mm_set_sd(b), _CMP_ORD_Q), _mm_set_sd(0.0)); #elif 0 return _mm_comineq_sd(_mm_set_sd(a), _mm_set_sd(a)) || _mm_comineq_sd(_mm_set_sd(b), _mm_set_sd(b)); #else return /* 1 & */ _mm_movemask_pd(_mm_cmp_sd(_mm_set_sd(a), _mm_set_sd(b), _CMP_UNORD_Q)); #endif } uint32_t lrint(double argument) { return _mm_cvtsd_si32(_mm_set_sd(argument)); } uint64_t llrint(double argument) { return _mm_cvtsd_si64(_mm_set_sd(argument)); } double copysign(double magnitude, double sign) { return _mm_cvtsd_f64(_mm_or_pd(_mm_and_pd(_mm_set_sd(-0.0), _mm_set_sd(sign)), _mm_andnot_pd(_mm_set_sd(-0.0), _mm_set_sd(magnitude)))); } double fdim(double x, double y) { return _mm_cvtsd_f64(_mm_and_pd(_mm_cmp_sd(_mm_set_sd(x), _mm_set_sd(y), _CMP_NLE_US), _mm_sub_sd(_mm_set_sd(x), _mm_set_sd(y)))); } double fmax(double x, double y) { __m128d mask = _mm_cmp_sd(_mm_set_sd(x), _mm_set_sd(x), _CMP_ORD_Q); return _mm_cvtsd_f64(_mm_or_pd(_mm_and_pd(mask, _mm_max_sd(_mm_set_sd(y), _mm_set_sd(x))), _mm_andnot_pd(mask, _mm_set_sd(y)))); } double fmin(double x, double y) { __m128d mask = _mm_cmp_sd(_mm_set_sd(x), _mm_set_sd(x), _CMP_ORD_Q); return _mm_cvtsd_f64(_mm_or_pd(_mm_and_pd(mask, _mm_min_sd(_mm_set_sd(y), _mm_set_sd(x))), _mm_andnot_pd(mask, _mm_set_sd(y)))); } Although the arguments and results are all held in SSE registers, there's no way to use them directly; it's but necessary to transfer them using _mm_set_sd() and _mm_cvtsd_f64(), which may result in superfluous instructions emitted by the compiler. If you but cheat and "hide" these functions from the compiler by placing them in a library, you can implement them as follows: __m128d fmin(__m128d x, __m128d y) { __m128d mask = _mm_cmp_sd(x, x, _CMP_ORD_Q); return _mm_or_pd(_mm_and_pd(mask, _mm_min_sd(y, x)), _mm_andnot_pd(mask, y)); } .code ; Intel syntax fmin proc public movsd xmm2, xmm0 ; xmm2 = x cmpsd xmm2, xmm0, 7 ; xmm2 = (x != NAN) ? -1 : 0 movsd xmm3, xmm2 andnpd xmm3, xmm1 ; xmm3 = (x != NAN) ? 0.0 : y minsd xmm1, xmm0 ; xmm1 = (x < y) ? x : y ; = min(x, y) andpd xmm2, xmm1 ; xmm2 = (x != NAN) ? min(x, y) : 0.0 orpd xmm2, xmm3 ; xmm2 = (x != NAN) ? min(x, y) : y movsd xmm0, xmm2 ; xmm0 = fmin(x, y) ret fmin endp
Powered by blists - more mailing lists
Confused about mailing lists and their use? Read about mailing lists on Wikipedia and check out these guidelines on proper formatting of your messages.