Lines Matching refs:z

264 	unsigned __int128 z, ff;  in f256_montymul()  local
281 z = (unsigned __int128)b[0] * x; in f256_montymul()
282 f = (uint64_t)z; in f256_montymul()
283 z = (unsigned __int128)b[1] * x + (z >> 64) + (uint64_t)(f << 32); in f256_montymul()
284 t0 = (uint64_t)z; in f256_montymul()
285 z = (unsigned __int128)b[2] * x + (z >> 64) + (uint64_t)(f >> 32); in f256_montymul()
286 t1 = (uint64_t)z; in f256_montymul()
287 z = (unsigned __int128)b[3] * x + (z >> 64) + f; in f256_montymul()
288 t2 = (uint64_t)z; in f256_montymul()
289 t3 = (uint64_t)(z >> 64); in f256_montymul()
291 z = (unsigned __int128)t2 + (uint64_t)ff; in f256_montymul()
292 t2 = (uint64_t)z; in f256_montymul()
293 z = (unsigned __int128)t3 + (z >> 64) + (ff >> 64); in f256_montymul()
294 t3 = (uint64_t)z; in f256_montymul()
295 t4 = (uint64_t)(z >> 64); in f256_montymul()
304 z = (unsigned __int128)b[0] * x + t0; in f256_montymul()
305 f = (uint64_t)z; in f256_montymul()
306 z = (unsigned __int128)b[1] * x + t1 + (z >> 64); in f256_montymul()
307 t0 = (uint64_t)z; in f256_montymul()
308 z = (unsigned __int128)b[2] * x + t2 + (z >> 64); in f256_montymul()
309 t1 = (uint64_t)z; in f256_montymul()
310 z = (unsigned __int128)b[3] * x + t3 + (z >> 64); in f256_montymul()
311 t2 = (uint64_t)z; in f256_montymul()
312 z = t4 + (z >> 64); in f256_montymul()
313 t3 = (uint64_t)z; in f256_montymul()
314 t4 = (uint64_t)(z >> 64); in f256_montymul()
317 z = (unsigned __int128)t0 + (uint64_t)(f << 32); in f256_montymul()
318 t0 = (uint64_t)z; in f256_montymul()
319 z = (z >> 64) + (unsigned __int128)t1 + (uint64_t)(f >> 32); in f256_montymul()
320 t1 = (uint64_t)z; in f256_montymul()
325 z = (z >> 64) + (unsigned __int128)t2 + (uint64_t)ff; in f256_montymul()
326 t2 = (uint64_t)z; in f256_montymul()
327 z = (unsigned __int128)t3 + (z >> 64) + (ff >> 64); in f256_montymul()
328 t3 = (uint64_t)z; in f256_montymul()
329 t4 += (uint64_t)(z >> 64); in f256_montymul()
350 z = (unsigned __int128)t0 + t4; in f256_montymul()
351 t0 = (uint64_t)z; in f256_montymul()
352 z = (unsigned __int128)t1 - (t4 << 32) + (z >> 64); in f256_montymul()
353 t1 = (uint64_t)z; in f256_montymul()
354 z = (unsigned __int128)t2 - (z >> 127); in f256_montymul()
355 t2 = (uint64_t)z; in f256_montymul()
356 t3 = t3 - (uint64_t)(z >> 127) - t4 + (t4 << 32); in f256_montymul()
601 unsigned __int128 z; in f256_final_reduce() local
608 z = (unsigned __int128)a[0] + 1; in f256_final_reduce()
609 t0 = (uint64_t)z; in f256_final_reduce()
610 z = (unsigned __int128)a[1] + (z >> 64) - ((uint64_t)1 << 32); in f256_final_reduce()
611 t1 = (uint64_t)z; in f256_final_reduce()
612 z = (unsigned __int128)a[2] - (z >> 127); in f256_final_reduce()
613 t2 = (uint64_t)z; in f256_final_reduce()
614 z = (unsigned __int128)a[3] - (z >> 127) + 0xFFFFFFFF; in f256_final_reduce()
615 t3 = (uint64_t)z; in f256_final_reduce()
616 cc = -(uint64_t)(z >> 64); in f256_final_reduce()
657 uint64_t z[4]; member
718 memcpy(P->z, F256_R, sizeof F256_R); in point_decode()
735 uint64_t t1[4], t2[4], z; in point_encode() local
738 f256_invert(t2, P->z); in point_encode()
765 z = P->z[0] | P->z[1] | P->z[2] | P->z[3]; in point_encode()
766 return NEQ((uint32_t)(z | z >> 32), 0); in point_encode()
803 f256_montysquare(t1, P->z); in p256_double()
836 f256_montymul(t4, P->y, P->z); in p256_double()
837 f256_add(P->z, t4, t4); in p256_double()
905 f256_montysquare(t3, P2->z); in p256_add()
907 f256_montymul(t4, P2->z, t3); in p256_add()
913 f256_montysquare(t4, P1->z); in p256_add()
915 f256_montymul(t5, P1->z, t4); in p256_add()
956 f256_montymul(t1, P1->z, P2->z); in p256_add()
957 f256_montymul(P1->z, t1, t2); in p256_add()
1023 f256_montysquare(t4, P1->z); in p256_add_mixed()
1025 f256_montymul(t5, P1->z, t4); in p256_add_mixed()
1066 f256_montymul(P1->z, P1->z, t2); in p256_add_mixed()
1134 zz = P1->z[0] | P1->z[1] | P1->z[2] | P1->z[3];
1146 f256_montysquare(t4, P1->z);
1148 f256_montymul(t5, P1->z, t4);
1193 f256_montymul(P1->z, P1->z, t2);
1243 P1->z[i] |= tt & t1[i];
1254 P1->z[i] ^= zz & (P1->z[i] ^ F256_R[i]);
1328 Q.z[j] |= m & F256_R[j]; in point_mul_inner()
1406 uint64_t z[16][4]; in window_to_affine() local
1408 #define zt (z[15]) in window_to_affine()
1409 #define zu (z[14]) in window_to_affine()
1410 #define zv (z[13]) in window_to_affine()
1418 memcpy(zt, jac[i].z, sizeof zt); in window_to_affine()
1419 memcpy(jac[i].z, jac[i + 1].z, sizeof zt); in window_to_affine()
1420 memcpy(jac[i + 1].z, zt, sizeof zt); in window_to_affine()
1421 f256_montymul(z[i >> 1], jac[i].z, jac[i + 1].z); in window_to_affine()
1424 memcpy(z[num >> 1], jac[num - 1].z, sizeof zt); in window_to_affine()
1425 memcpy(jac[num - 1].z, F256_R, sizeof F256_R); in window_to_affine()
1437 f256_montymul(jac[i].z, jac[i].z, z[(i >> k) ^ 1]); in window_to_affine()
1441 f256_montymul(z[i], z[i << 1], z[(i << 1) + 1]); in window_to_affine()
1444 memmove(z[n >> 1], z[n], sizeof zt); in window_to_affine()
1451 f256_invert(zt, z[0]); in window_to_affine()
1453 f256_montymul(zv, jac[i].z, zt); in window_to_affine()
1628 uint32_t z; in check_scalar() local
1635 z = 0; in check_scalar()
1637 z |= k[u]; in check_scalar()
1647 return NEQ(z, 0) & LT0(c); in check_scalar()
1715 uint64_t z; in api_muladd() local
1734 f256_final_reduce(P.z); in api_muladd()
1735 z = P.z[0] | P.z[1] | P.z[2] | P.z[3]; in api_muladd()
1736 s = EQ((uint32_t)(z | (z >> 32)), 0); in api_muladd()