| /**************************************************************************** |
| * crypto/ecc.c |
| * |
| * SPDX-License-Identifier: BSD-2-Clause |
| * |
| * Copyright (c) 2013, Kenneth MacKay All rights reserved. |
| * |
| * Redistribution and use in source and binary forms, with or without |
| * modification, are permitted provided that the following conditions are |
| * met: Redistributions of source code must retain the above copyright |
| * notice, this list of conditions and the following disclaimer. |
| * Redistributions in binary form must reproduce the above copyright notice, |
| * this list of conditions and the following disclaimer in the documentation |
| * and/or other materials provided with the distribution. |
| |
| * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS |
| * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT |
| * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR |
| * A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT |
| * SPECIAL, HOLDER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, |
| * INCIDENTAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT |
| * LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, |
| * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY |
| * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT |
| * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE |
| * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. |
| ****************************************************************************/ |
| |
| /**************************************************************************** |
| * Included Files |
| ****************************************************************************/ |
| |
| #include <fcntl.h> |
| #include <stdlib.h> |
| #include <string.h> |
| #include <unistd.h> |
| #include <sys/types.h> |
| |
| #include <crypto/ecc.h> |
| #include <nuttx/macro.h> |
| |
| /**************************************************************************** |
| * Pre-processor Definitions |
| ****************************************************************************/ |
| |
| #define NUM_ECC_DIGITS (ECC_BYTES / 8) |
| #define MAX_TRIES 16 |
| |
| #define EVEN(vli) (!(vli[0] & 1)) |
| |
| #define curve_p_16 { 0xFFFFFFFFFFFFFFFF, 0xFFFFFFFDFFFFFFFF } |
| #define curve_p_24 { 0xFFFFFFFFFFFFFFFFull, 0xFFFFFFFFFFFFFFFEull, \ |
| 0xFFFFFFFFFFFFFFFFull } |
| #define curve_p_32 { 0xFFFFFFFFFFFFFFFFull, 0x00000000FFFFFFFFull, \ |
| 0x0000000000000000ull, 0xFFFFFFFF00000001ull } |
| #define curve_p_48 { 0x00000000FFFFFFFF, 0xFFFFFFFF00000000, \ |
| 0xFFFFFFFFFFFFFFFE, 0xFFFFFFFFFFFFFFFF, \ |
| 0xFFFFFFFFFFFFFFFF, 0xFFFFFFFFFFFFFFFF } |
| |
| #define curve_b_16 { 0xD824993C2CEE5ED3, 0xE87579C11079F43D } |
| #define curve_b_24 { 0xFEB8DEECC146B9B1ull, 0x0FA7E9AB72243049ull, \ |
| 0x64210519E59C80E7ull } |
| #define curve_b_32 { 0x3BCE3C3E27D2604Bull, 0x651D06B0CC53B0F6ull, \ |
| 0xB3EBBD55769886BCull, 0x5AC635D8AA3A93E7ull } |
| #define curve_b_48 { 0x2A85C8EDD3EC2AEF, 0xC656398D8A2ED19D, \ |
| 0x0314088F5013875A, 0x181D9C6EFE814112, \ |
| 0x988E056BE3F82D19, 0xB3312FA7E23EE7E4 } |
| |
| #define curve_g_16 { \ |
| { 0x0C28607CA52C5B86, 0x161FF7528B899B2D }, \ |
| { 0xC02DA292DDED7A83, 0xCF5AC8395BAFEB13 }} |
| |
| #define curve_g_24 { \ |
| { 0xF4FF0AFD82FF1012ull, 0x7CBF20EB43A18800ull, 0x188DA80EB03090F6ull }, \ |
| { 0x73F977A11E794811ull, 0x631011ED6B24CDD5ull, 0x07192B95FFC8DA78ull }} |
| |
| #define curve_g_32 { \ |
| { 0xF4A13945D898C296ull, 0x77037D812DEB33A0ull, \ |
| 0xF8BCE6E563A440F2ull, 0x6B17D1F2E12C4247ull }, \ |
| { 0xCBB6406837BF51F5ull, 0x2BCE33576B315ECEull, \ |
| 0x8EE7EB4A7C0F9E16ull, 0x4FE342E2FE1A7F9Bull }} |
| |
| #define curve_g_48 { \ |
| { 0x3A545E3872760AB7, 0x5502F25DBF55296C, 0x59F741E082542A38, \ |
| 0x6E1D3B628BA79B98, 0x8EB1C71EF320AD74, 0xAA87CA22BE8B0537}, \ |
| { 0x7A431D7C90EA0E5F, 0x0A60B1CE1D7E819D, 0xE9DA3113B5F0B8C0, \ |
| 0xF8F41DBD289A147C, 0x5D9E98BF9292DC29, 0x3617DE4A96262C6F }} |
| |
| #define curve_n_16 { 0x75A30D1B9038A115, 0xFFFFFFFE00000000 } |
| #define curve_n_24 { 0x146BC9B1B4D22831ull, 0xFFFFFFFF99DEF836ull, \ |
| 0xFFFFFFFFFFFFFFFFull } |
| #define curve_n_32 { 0xF3B9CAC2FC632551ull, 0xBCE6FAADA7179E84ull, \ |
| 0xFFFFFFFFFFFFFFFFull, 0xFFFFFFFF00000000ull } |
| #define curve_n_48 { 0xECEC196ACCC52973, 0x581A0DB248B0A77A, \ |
| 0xC7634D81F4372DDF, 0xFFFFFFFFFFFFFFFF, \ |
| 0xFFFFFFFFFFFFFFFF, 0xFFFFFFFFFFFFFFFF } |
| |
| #if defined(__SIZEOF_INT128__) |
| # define SUPPORTS_INT128 1 |
| #else |
| # define SUPPORTS_INT128 0 |
| #endif |
| |
| /**************************************************************************** |
| * Private Type Definitions |
| ****************************************************************************/ |
| |
| #if SUPPORTS_INT128 |
| typedef unsigned __int128 uint128_t; |
| #else |
| typedef struct |
| { |
| uint64_t m_low; |
| uint64_t m_high; |
| } uint128_t; |
| #endif |
| |
| typedef struct |
| { |
| uint64_t x[NUM_ECC_DIGITS]; |
| uint64_t y[NUM_ECC_DIGITS]; |
| } eccpoint_t; |
| |
| /**************************************************************************** |
| * Private Data |
| ****************************************************************************/ |
| |
| static uint64_t g_curve_p[NUM_ECC_DIGITS] = CONCATENATE(curve_p_, ECC_CURVE); |
| static uint64_t g_curve_b[NUM_ECC_DIGITS] = CONCATENATE(curve_b_, ECC_CURVE); |
| static uint64_t g_curve_n[NUM_ECC_DIGITS] = CONCATENATE(curve_n_, ECC_CURVE); |
| static eccpoint_t g_curve_g = CONCATENATE(curve_g_, ECC_CURVE); |
| |
| /**************************************************************************** |
| * Private Functions |
| ****************************************************************************/ |
| |
| static void vli_clear(FAR uint64_t *vli) |
| { |
| uint i; |
| |
| for (i = 0; i < NUM_ECC_DIGITS; ++i) |
| { |
| vli[i] = 0; |
| } |
| } |
| |
| /* Returns 1 if vli == 0, 0 otherwise. */ |
| |
| static int vli_iszero(FAR uint64_t *vli) |
| { |
| uint i; |
| |
| for (i = 0; i < NUM_ECC_DIGITS; ++i) |
| { |
| if (vli[i]) |
| { |
| return 0; |
| } |
| } |
| |
| return 1; |
| } |
| |
| /* Returns nonzero if bit bit of vli is set. */ |
| |
| static uint64_t vli_testbit(FAR uint64_t *vli, uint bit) |
| { |
| return vli[bit / 64] & ((uint64_t)1 << (bit % 64)); |
| } |
| |
| /* Counts the number of 64-bit "digits" in vli. */ |
| |
| static uint vli_numdigits(FAR uint64_t *vli) |
| { |
| int i; |
| |
| /* Search from the end until we find a non-zero digit. |
| * We do it in reverse because we expect that most digits |
| * will be nonzero. |
| */ |
| |
| for (i = NUM_ECC_DIGITS - 1; i >= 0 && vli[i] == 0; --i) |
| { |
| } |
| |
| return i + 1; |
| } |
| |
| /* Counts the number of bits required for vli. */ |
| |
| static uint vli_numbits(FAR uint64_t *vli) |
| { |
| uint64_t l_digit; |
| uint l_numdigits = vli_numdigits(vli); |
| uint i; |
| |
| if (l_numdigits == 0) |
| { |
| return 0; |
| } |
| |
| l_digit = vli[l_numdigits - 1]; |
| for (i = 0; l_digit; ++i) |
| { |
| l_digit >>= 1; |
| } |
| |
| return (l_numdigits - 1) * 64 + i; |
| } |
| |
| /* Sets dest = src. */ |
| |
| static void vli_set(FAR uint64_t *dest, FAR uint64_t *src) |
| { |
| uint i; |
| |
| for (i = 0; i < NUM_ECC_DIGITS; ++i) |
| { |
| dest[i] = src[i]; |
| } |
| } |
| |
| /* Returns sign of left - right. */ |
| |
| static int vli_cmp(FAR uint64_t *left, FAR uint64_t *right) |
| { |
| int i; |
| |
| for (i = NUM_ECC_DIGITS - 1; i >= 0; --i) |
| { |
| if (left[i] > right[i]) |
| { |
| return 1; |
| } |
| else if (left[i] < right[i]) |
| { |
| return -1; |
| } |
| } |
| |
| return 0; |
| } |
| |
| /* Computes result = in << c, returning carry. |
| * Can modify in place (if result == in). 0 < shift < 64. |
| */ |
| |
| static uint64_t vli_lshift(FAR uint64_t *result, FAR uint64_t *in, |
| uint shift) |
| { |
| uint64_t l_carry = 0; |
| uint64_t l_temp; |
| uint i; |
| |
| for (i = 0; i < NUM_ECC_DIGITS; ++i) |
| { |
| l_temp = in[i]; |
| result[i] = (l_temp << shift) | l_carry; |
| l_carry = l_temp >> (64 - shift); |
| } |
| |
| return l_carry; |
| } |
| |
| /* Computes vli = vli >> 1. */ |
| |
| static void vli_rshift1(FAR uint64_t *vli) |
| { |
| FAR uint64_t *l_end = vli; |
| uint64_t l_carry = 0; |
| uint64_t l_temp; |
| |
| vli += NUM_ECC_DIGITS; |
| while (vli-- > l_end) |
| { |
| l_temp = *vli; |
| *vli = (l_temp >> 1) | l_carry; |
| l_carry = l_temp << 63; |
| } |
| } |
| |
| /* Computes result = left + right, returning carry. Can modify in place. */ |
| |
| static uint64_t vli_add(FAR uint64_t *result, FAR uint64_t *left, |
| FAR uint64_t *right) |
| { |
| uint64_t l_carry = 0; |
| uint64_t l_sum; |
| uint i; |
| |
| for (i = 0; i < NUM_ECC_DIGITS; ++i) |
| { |
| l_sum = left[i] + right[i] + l_carry; |
| if (l_sum != left[i]) |
| { |
| l_carry = (l_sum < left[i]); |
| } |
| |
| result[i] = l_sum; |
| } |
| |
| return l_carry; |
| } |
| |
| /* Computes result = left - right, returning borrow. Can modify in place. */ |
| |
| static uint64_t vli_sub(FAR uint64_t *result, FAR uint64_t *left, |
| FAR uint64_t *right) |
| { |
| uint64_t l_borrow = 0; |
| uint64_t l_diff; |
| uint i; |
| |
| for (i = 0; i < NUM_ECC_DIGITS; ++i) |
| { |
| l_diff = left[i] - right[i] - l_borrow; |
| if (l_diff != left[i]) |
| { |
| l_borrow = (l_diff > left[i]); |
| } |
| |
| result[i] = l_diff; |
| } |
| |
| return l_borrow; |
| } |
| |
| #if SUPPORTS_INT128 |
| |
| /* Computes result = left * right. */ |
| |
| static void vli_mult(FAR uint64_t *result, FAR uint64_t *left, |
| FAR uint64_t *right) |
| { |
| uint128_t l_product; |
| uint128_t r01 = 0; |
| uint64_t r2 = 0; |
| uint l_min; |
| uint i; |
| uint k; |
| |
| /* Compute each digit of result in sequence, maintaining the carries. */ |
| |
| for (k = 0; k < NUM_ECC_DIGITS * 2 - 1; ++k) |
| { |
| l_min = (k < NUM_ECC_DIGITS ? 0 : (k + 1) - NUM_ECC_DIGITS); |
| for (i = l_min; i <= k && i < NUM_ECC_DIGITS; ++i) |
| { |
| l_product = (uint128_t)left[i] * right[k - i]; |
| r01 += l_product; |
| r2 += (r01 < l_product); |
| } |
| |
| result[k] = (uint64_t)r01; |
| r01 = (r01 >> 64) | (((uint128_t)r2) << 64); |
| r2 = 0; |
| } |
| |
| result[NUM_ECC_DIGITS * 2 - 1] = (uint64_t)r01; |
| } |
| |
| /* Computes result = left^2. */ |
| |
| static void vli_square(FAR uint64_t *result, FAR uint64_t *left) |
| { |
| uint128_t l_product; |
| uint128_t r01 = 0; |
| uint64_t r2 = 0; |
| uint l_min; |
| uint i; |
| uint k; |
| |
| for (k = 0; k < NUM_ECC_DIGITS * 2 - 1; ++k) |
| { |
| l_min = (k < NUM_ECC_DIGITS ? 0 : (k + 1) - NUM_ECC_DIGITS); |
| for (i = l_min; i <= k && i <= k - i; ++i) |
| { |
| l_product = (uint128_t)left[i] * left[k - i]; |
| if (i < k - i) |
| { |
| r2 += l_product >> 127; |
| l_product *= 2; |
| } |
| |
| r01 += l_product; |
| r2 += (r01 < l_product); |
| } |
| |
| result[k] = (uint64_t)r01; |
| r01 = (r01 >> 64) | (((uint128_t)r2) << 64); |
| r2 = 0; |
| } |
| |
| result[NUM_ECC_DIGITS * 2 - 1] = (uint64_t)r01; |
| } |
| |
| #else /* #if SUPPORTS_INT128 */ |
| |
| static uint128_t mul_64_64(uint64_t left, uint64_t right) |
| { |
| uint128_t l_result; |
| uint64_t a0 = left & 0xffffffffull; |
| uint64_t a1 = left >> 32; |
| uint64_t b0 = right & 0xffffffffull; |
| uint64_t b1 = right >> 32; |
| uint64_t m0 = a0 * b0; |
| uint64_t m1 = a0 * b1; |
| uint64_t m2 = a1 * b0; |
| uint64_t m3 = a1 * b1; |
| |
| m2 += (m0 >> 32); |
| m2 += m1; |
| if (m2 < m1) |
| { |
| m3 += 0x100000000ull; |
| } |
| |
| l_result.m_low = (m0 & 0xffffffffull) | (m2 << 32); |
| l_result.m_high = m3 + (m2 >> 32); |
| |
| return l_result; |
| } |
| |
| static uint128_t add_128_128(uint128_t a, uint128_t b) |
| { |
| uint128_t l_result; |
| |
| l_result.m_low = a.m_low + b.m_low; |
| l_result.m_high = a.m_high + b.m_high + (l_result.m_low < a.m_low); |
| return l_result; |
| } |
| |
| static void vli_mult(FAR uint64_t *result, FAR uint64_t *left, |
| FAR uint64_t *right) |
| { |
| uint64_t r2 = 0; |
| uint i; |
| uint k; |
| uint128_t l_product; |
| uint128_t r01 = |
| { |
| 0, 0 |
| }; |
| |
| /* Compute each digit of result in sequence, maintaining the carries. */ |
| |
| for (k = 0; k < NUM_ECC_DIGITS * 2 - 1; ++k) |
| { |
| uint l_min = (k < NUM_ECC_DIGITS ? 0 : (k + 1) - NUM_ECC_DIGITS); |
| for (i = l_min; i <= k && i < NUM_ECC_DIGITS; ++i) |
| { |
| l_product = mul_64_64(left[i], right[k - i]); |
| r01 = add_128_128(r01, l_product); |
| r2 += (r01.m_high < l_product.m_high); |
| } |
| |
| result[k] = r01.m_low; |
| r01.m_low = r01.m_high; |
| r01.m_high = r2; |
| r2 = 0; |
| } |
| |
| result[NUM_ECC_DIGITS * 2 - 1] = r01.m_low; |
| } |
| |
| static void vli_square(FAR uint64_t *result, FAR uint64_t *left) |
| { |
| uint64_t r2 = 0; |
| uint l_min; |
| uint i; |
| uint k; |
| uint128_t l_product; |
| uint128_t r01 = |
| { |
| 0, 0 |
| }; |
| |
| for (k = 0; k < NUM_ECC_DIGITS * 2 - 1; ++k) |
| { |
| l_min = (k < NUM_ECC_DIGITS ? 0 : (k + 1) - NUM_ECC_DIGITS); |
| for (i = l_min; i <= k && i <= k - i; ++i) |
| { |
| l_product = mul_64_64(left[i], left[k - i]); |
| if (i < k - i) |
| { |
| r2 += l_product.m_high >> 63; |
| l_product.m_high = (l_product.m_high << 1) | |
| (l_product.m_low >> 63); |
| l_product.m_low <<= 1; |
| } |
| |
| r01 = add_128_128(r01, l_product); |
| r2 += (r01.m_high < l_product.m_high); |
| } |
| |
| result[k] = r01.m_low; |
| r01.m_low = r01.m_high; |
| r01.m_high = r2; |
| r2 = 0; |
| } |
| |
| result[NUM_ECC_DIGITS * 2 - 1] = r01.m_low; |
| } |
| |
| #endif /* SUPPORTS_INT128 */ |
| |
| /* Computes result = (left + right) % mod. |
| * Assumes that left < mod and right < mod, result != mod. |
| */ |
| |
| static void vli_modadd(FAR uint64_t *result, FAR uint64_t *left, |
| FAR uint64_t *right, FAR uint64_t *mod) |
| { |
| uint64_t l_carry = vli_add(result, left, right); |
| |
| if (l_carry || vli_cmp(result, mod) >= 0) |
| { |
| /* result > mod (result = mod + remainder), |
| * so subtract mod to get remainder. |
| */ |
| |
| vli_sub(result, result, mod); |
| } |
| } |
| |
| /* Computes result = (left - right) % mod. |
| * Assumes that left < mod and right < mod, result != mod. |
| */ |
| |
| static void vli_modsub(FAR uint64_t *result, FAR uint64_t *left, |
| FAR uint64_t *right, FAR uint64_t *mod) |
| { |
| uint64_t l_borrow = vli_sub(result, left, right); |
| |
| if (l_borrow) |
| { |
| /* In this case, result == -diff == (max int) - diff. |
| * Since -x % d == d - x, we can get the correct result from |
| * result + mod (with overflow). |
| */ |
| |
| vli_add(result, result, mod); |
| } |
| } |
| |
| #if ECC_CURVE == secp128r1 |
| |
| /* Computes result = product % g_curve_p. |
| * See algorithm 5 and 6 from |
| * http://www.isys.uni-klu.ac.at/PDF/2001-0126-MT.pdf |
| */ |
| |
| static void vli_mmod_fast(FAR uint64_t *result, |
| FAR uint64_t *product) |
| { |
| uint64_t l_tmp[NUM_ECC_DIGITS]; |
| int l_carry; |
| |
| vli_set(result, product); |
| |
| l_tmp[0] = product[2]; |
| l_tmp[1] = (product[3] & 0x1ffffffffull) | (product[2] << 33); |
| l_carry = vli_add(result, result, l_tmp); |
| |
| l_tmp[0] = (product[2] >> 31) | (product[3] << 33); |
| l_tmp[1] = (product[3] >> 31) | |
| ((product[2] & 0xffffffff80000000ull) << 2); |
| l_carry += vli_add(result, result, l_tmp); |
| |
| l_tmp[0] = (product[2] >> 62) | (product[3] << 2); |
| l_tmp[1] = (product[3] >> 62) | |
| ((product[2] & 0xc000000000000000ull) >> 29) | |
| (product[3] << 35); |
| l_carry += vli_add(result, result, l_tmp); |
| |
| l_tmp[0] = (product[3] >> 29); |
| l_tmp[1] = ((product[3] & 0xffffffffe0000000ull) << 4); |
| l_carry += vli_add(result, result, l_tmp); |
| |
| l_tmp[0] = (product[3] >> 60); |
| l_tmp[1] = (product[3] & 0xfffffffe00000000ull); |
| l_carry += vli_add(result, result, l_tmp); |
| |
| l_tmp[0] = 0; |
| l_tmp[1] = ((product[3] & 0xf000000000000000ull) >> 27); |
| l_carry += vli_add(result, result, l_tmp); |
| |
| while (l_carry || vli_cmp(g_curve_p, result) != 1) |
| { |
| l_carry -= vli_sub(result, result, g_curve_p); |
| } |
| } |
| |
| #elif ECC_CURVE == secp192r1 |
| |
| /* Computes result = product % g_curve_p. |
| * See algorithm 5 and 6 from |
| * http://www.isys.uni-klu.ac.at/PDF/2001-0126-MT.pdf |
| */ |
| |
| static void vli_mmod_fast(FAR uint64_t *result, |
| FAR uint64_t *product) |
| { |
| uint64_t l_tmp[NUM_ECC_DIGITS]; |
| int l_carry; |
| |
| vli_set(result, product); |
| |
| vli_set(l_tmp, &product[3]); |
| l_carry = vli_add(result, result, l_tmp); |
| |
| l_tmp[0] = 0; |
| l_tmp[1] = product[3]; |
| l_tmp[2] = product[4]; |
| l_carry += vli_add(result, result, l_tmp); |
| |
| l_tmp[0] = l_tmp[1] = product[5]; |
| l_tmp[2] = 0; |
| l_carry += vli_add(result, result, l_tmp); |
| |
| while (l_carry || vli_cmp(g_curve_p, result) != 1) |
| { |
| l_carry -= vli_sub(result, result, g_curve_p); |
| } |
| } |
| |
| #elif ECC_CURVE == secp256r1 |
| |
| /* Computes result = product % g_curve_p |
| * from http://www.nsa.gov/ia/_files/nist-routines.pdf |
| */ |
| |
| static void vli_mmod_fast(FAR uint64_t *result, |
| FAR uint64_t *product) |
| { |
| uint64_t l_tmp[NUM_ECC_DIGITS]; |
| int l_carry; |
| |
| /* t */ |
| |
| vli_set(result, product); |
| |
| /* s1 */ |
| |
| l_tmp[0] = 0; |
| l_tmp[1] = product[5] & 0xffffffff00000000ull; |
| l_tmp[2] = product[6]; |
| l_tmp[3] = product[7]; |
| l_carry = vli_lshift(l_tmp, l_tmp, 1); |
| l_carry += vli_add(result, result, l_tmp); |
| |
| /* s2 */ |
| |
| l_tmp[1] = product[6] << 32; |
| l_tmp[2] = (product[6] >> 32) | (product[7] << 32); |
| l_tmp[3] = product[7] >> 32; |
| l_carry += vli_lshift(l_tmp, l_tmp, 1); |
| l_carry += vli_add(result, result, l_tmp); |
| |
| /* s3 */ |
| |
| l_tmp[0] = product[4]; |
| l_tmp[1] = product[5] & 0xffffffff; |
| l_tmp[2] = 0; |
| l_tmp[3] = product[7]; |
| l_carry += vli_add(result, result, l_tmp); |
| |
| /* s4 */ |
| |
| l_tmp[0] = (product[4] >> 32) | (product[5] << 32); |
| l_tmp[1] = (product[5] >> 32) | (product[6] & 0xffffffff00000000ull); |
| l_tmp[2] = product[7]; |
| l_tmp[3] = (product[6] >> 32) | (product[4] << 32); |
| l_carry += vli_add(result, result, l_tmp); |
| |
| /* d1 */ |
| |
| l_tmp[0] = (product[5] >> 32) | (product[6] << 32); |
| l_tmp[1] = (product[6] >> 32); |
| l_tmp[2] = 0; |
| l_tmp[3] = (product[4] & 0xffffffff) | (product[5] << 32); |
| l_carry -= vli_sub(result, result, l_tmp); |
| |
| /* d2 */ |
| |
| l_tmp[0] = product[6]; |
| l_tmp[1] = product[7]; |
| l_tmp[2] = 0; |
| l_tmp[3] = (product[4] >> 32) | (product[5] & 0xffffffff00000000ull); |
| l_carry -= vli_sub(result, result, l_tmp); |
| |
| /* d3 */ |
| |
| l_tmp[0] = (product[6] >> 32) | (product[7] << 32); |
| l_tmp[1] = (product[7] >> 32) | (product[4] << 32); |
| l_tmp[2] = (product[4] >> 32) | (product[5] << 32); |
| l_tmp[3] = (product[6] << 32); |
| l_carry -= vli_sub(result, result, l_tmp); |
| |
| /* d4 */ |
| |
| l_tmp[0] = product[7]; |
| l_tmp[1] = product[4] & 0xffffffff00000000ull; |
| l_tmp[2] = product[5]; |
| l_tmp[3] = product[6] & 0xffffffff00000000ull; |
| l_carry -= vli_sub(result, result, l_tmp); |
| |
| if (l_carry < 0) |
| { |
| do |
| { |
| l_carry += vli_add(result, result, g_curve_p); |
| } |
| while (l_carry < 0); |
| } |
| else |
| { |
| while (l_carry || vli_cmp(g_curve_p, result) != 1) |
| { |
| l_carry -= vli_sub(result, result, g_curve_p); |
| } |
| } |
| } |
| |
| #elif ECC_CURVE == secp384r1 |
| |
| static void omega_mult(uint64_t *result, uint64_t *right) |
| { |
| uint64_t l_tmp[NUM_ECC_DIGITS]; |
| uint64_t l_carry; |
| uint64_t l_diff; |
| uint i; |
| |
| /* Multiply by (2^128 + 2^96 - 2^32 + 1). */ |
| |
| vli_set(result, right); /* 1 */ |
| l_carry = vli_lshift(l_tmp, right, 32); |
| result[1 + NUM_ECC_DIGITS] = l_carry + vli_add(result + 1, result + 1, l_tmp); /* 2^96 + 1 */ |
| |
| /* 2^128 + 2^96 + 1 */ |
| |
| result[2 + NUM_ECC_DIGITS] = vli_add(result + 2, result + 2, right); |
| |
| /* 2^128 + 2^96 - 2^32 + 1 */ |
| |
| l_carry += vli_sub(result, result, l_tmp); |
| l_diff = result[NUM_ECC_DIGITS] - l_carry; |
| if (l_diff > result[NUM_ECC_DIGITS]) |
| { |
| /* Propagate borrow if necessary. */ |
| |
| for (i = 1 + NUM_ECC_DIGITS; ; ++i) |
| { |
| --result[i]; |
| if (result[i] != (uint64_t)-1) |
| { |
| break; |
| } |
| } |
| } |
| |
| result[NUM_ECC_DIGITS] = l_diff; |
| } |
| |
| /* Computes result = product % g_curve_p |
| * see PDF "Comparing Elliptic Curve Cryptography and RSA on 8-bit CPUs" |
| * section "Curve-Specific Optimizations" |
| */ |
| |
| static void vli_mmod_fast(uint64_t *result, uint64_t *product) |
| { |
| uint64_t l_tmp[2 * NUM_ECC_DIGITS]; |
| uint64_t l_carry; |
| uint64_t l_sum; |
| uint i; |
| |
| while (!vli_iszero(product + NUM_ECC_DIGITS)) /* While c1 != 0 */ |
| { |
| l_carry = 0; |
| |
| vli_clear(l_tmp); |
| vli_clear(l_tmp + NUM_ECC_DIGITS); |
| omega_mult(l_tmp, product + NUM_ECC_DIGITS); /* tmp = w * c1 */ |
| |
| /* p = c0 */ |
| |
| vli_clear(product + NUM_ECC_DIGITS); |
| |
| /* (c1, c0) = c0 + w * c1 */ |
| |
| for (i = 0; i < NUM_ECC_DIGITS + 3; ++i) |
| { |
| l_sum = product[i] + l_tmp[i] + l_carry; |
| if (l_sum != product[i]) |
| { |
| l_carry = (l_sum < product[i]); |
| } |
| |
| product[i] = l_sum; |
| } |
| } |
| |
| while (vli_cmp(product, g_curve_p) > 0) |
| { |
| vli_sub(product, product, g_curve_p); |
| } |
| |
| vli_set(result, product); |
| } |
| |
| #endif |
| |
| /* Computes result = (left * right) % g_curve_p. */ |
| |
| static void vli_modmult_fast(FAR uint64_t *result, FAR uint64_t *left, |
| FAR uint64_t *right) |
| { |
| uint64_t l_product[2 * NUM_ECC_DIGITS]; |
| |
| vli_mult(l_product, left, right); |
| vli_mmod_fast(result, l_product); |
| } |
| |
| /* Computes result = left^2 % g_curve_p. */ |
| |
| static void vli_modsquare_fast(FAR uint64_t *result, |
| FAR uint64_t *left) |
| { |
| uint64_t l_product[2 * NUM_ECC_DIGITS]; |
| |
| vli_square(l_product, left); |
| vli_mmod_fast(result, l_product); |
| } |
| |
| /* Computes result = (1 / input) % mod. All VLIs are the same size. |
| * See "From Euclid's GCD to Montgomery Multiplication to the Great Divide" |
| * https://labs.oracle.com/techrep/2001/smli_tr-2001-95.pdf |
| */ |
| |
| static void vli_modinv(FAR uint64_t *result, FAR uint64_t *input, |
| FAR uint64_t *mod) |
| { |
| uint64_t a[NUM_ECC_DIGITS]; |
| uint64_t b[NUM_ECC_DIGITS]; |
| uint64_t u[NUM_ECC_DIGITS]; |
| uint64_t v[NUM_ECC_DIGITS]; |
| uint64_t l_carry; |
| int l_cmpresult; |
| |
| if (vli_iszero(input)) |
| { |
| vli_clear(result); |
| return; |
| } |
| |
| vli_set(a, input); |
| vli_set(b, mod); |
| vli_clear(u); |
| u[0] = 1; |
| vli_clear(v); |
| |
| while ((l_cmpresult = vli_cmp(a, b)) != 0) |
| { |
| l_carry = 0; |
| if (EVEN(a)) |
| { |
| vli_rshift1(a); |
| if (!EVEN(u)) |
| { |
| l_carry = vli_add(u, u, mod); |
| } |
| |
| vli_rshift1(u); |
| if (l_carry) |
| { |
| u[NUM_ECC_DIGITS - 1] |= 0x8000000000000000ull; |
| } |
| } |
| else if (EVEN(b)) |
| { |
| vli_rshift1(b); |
| if (!EVEN(v)) |
| { |
| l_carry = vli_add(v, v, mod); |
| } |
| |
| vli_rshift1(v); |
| if (l_carry) |
| { |
| v[NUM_ECC_DIGITS - 1] |= 0x8000000000000000ull; |
| } |
| } |
| else if (l_cmpresult > 0) |
| { |
| vli_sub(a, a, b); |
| vli_rshift1(a); |
| if (vli_cmp(u, v) < 0) |
| { |
| vli_add(u, u, mod); |
| } |
| |
| vli_sub(u, u, v); |
| if (!EVEN(u)) |
| { |
| l_carry = vli_add(u, u, mod); |
| } |
| |
| vli_rshift1(u); |
| if (l_carry) |
| { |
| u[NUM_ECC_DIGITS - 1] |= 0x8000000000000000ull; |
| } |
| } |
| else |
| { |
| vli_sub(b, b, a); |
| vli_rshift1(b); |
| if (vli_cmp(v, u) < 0) |
| { |
| vli_add(v, v, mod); |
| } |
| |
| vli_sub(v, v, u); |
| if (!EVEN(v)) |
| { |
| l_carry = vli_add(v, v, mod); |
| } |
| |
| vli_rshift1(v); |
| if (l_carry) |
| { |
| v[NUM_ECC_DIGITS - 1] |= 0x8000000000000000ull; |
| } |
| } |
| } |
| |
| vli_set(result, u); |
| } |
| |
| /* ------ Point operations ------ */ |
| |
| /* Returns 1 if point is the point at infinity, 0 otherwise. */ |
| |
| static int eccpoint_iszero(FAR eccpoint_t *point) |
| { |
| return vli_iszero(point->x) && vli_iszero(point->y); |
| } |
| |
| /* Point multiplication algorithm using Montgomery's ladder with |
| * co-Z coordinates. From http://eprint.iacr.org/2011/338.pdf |
| */ |
| |
| /* Double in place */ |
| |
| static void eccpoint_double_jacobian(FAR uint64_t *X1, |
| FAR uint64_t *Y1, |
| FAR uint64_t *Z1) |
| { |
| /* t1 = X, t2 = Y, t3 = Z */ |
| |
| uint64_t t4[NUM_ECC_DIGITS]; |
| uint64_t t5[NUM_ECC_DIGITS]; |
| uint64_t l_carry; |
| |
| if (vli_iszero(Z1)) |
| { |
| return; |
| } |
| |
| vli_modsquare_fast(t4, Y1); /* t4 = y1^2 */ |
| vli_modmult_fast(t5, X1, t4); /* t5 = x1*y1^2 = A */ |
| vli_modsquare_fast(t4, t4); /* t4 = y1^4 */ |
| vli_modmult_fast(Y1, Y1, Z1); /* t2 = y1*z1 = z3 */ |
| vli_modsquare_fast(Z1, Z1); /* t3 = z1^2 */ |
| |
| vli_modadd(X1, X1, Z1, g_curve_p); /* t1 = x1 + z1^2 */ |
| vli_modadd(Z1, Z1, Z1, g_curve_p); /* t3 = 2*z1^2 */ |
| vli_modsub(Z1, X1, Z1, g_curve_p); /* t3 = x1 - z1^2 */ |
| |
| /* t1 = x1^2 - z1^4 */ |
| |
| vli_modmult_fast(X1, X1, Z1); |
| |
| vli_modadd(Z1, X1, X1, g_curve_p); /* t3 = 2*(x1^2 - z1^4) */ |
| vli_modadd(X1, X1, Z1, g_curve_p); /* t1 = 3*(x1^2 - z1^4) */ |
| if (vli_testbit(X1, 0)) |
| { |
| l_carry = vli_add(X1, X1, g_curve_p); |
| vli_rshift1(X1); |
| X1[NUM_ECC_DIGITS - 1] |= l_carry << 63; |
| } |
| else |
| { |
| vli_rshift1(X1); |
| } |
| |
| /* t1 = 3/2*(x1^2 - z1^4) = B */ |
| |
| /* t3 = B^2 */ |
| |
| vli_modsquare_fast(Z1, X1); |
| |
| /* t3 = B^2 - A */ |
| |
| vli_modsub(Z1, Z1, t5, g_curve_p); |
| |
| /* t3 = B^2 - 2A = x3 */ |
| |
| vli_modsub(Z1, Z1, t5, g_curve_p); |
| |
| /* t5 = A - x3 */ |
| |
| vli_modsub(t5, t5, Z1, g_curve_p); |
| |
| /* t1 = B * (A - x3) */ |
| |
| vli_modmult_fast(X1, X1, t5); |
| |
| /* t4 = B * (A - x3) - y1^4 = y3 */ |
| |
| vli_modsub(t4, X1, t4, g_curve_p); |
| |
| vli_set(X1, Z1); |
| vli_set(Z1, Y1); |
| vli_set(Y1, t4); |
| } |
| |
| /* Modify (x1, y1) => (x1 * z^2, y1 * z^3) */ |
| |
| static void apply_z(FAR uint64_t *X1, FAR uint64_t *Y1, |
| FAR uint64_t *Z) |
| { |
| uint64_t t1[NUM_ECC_DIGITS]; |
| |
| /* z^2 */ |
| |
| vli_modsquare_fast(t1, Z); |
| |
| /* x1 * z^2 */ |
| |
| vli_modmult_fast(X1, X1, t1); |
| |
| /* z^3 */ |
| |
| vli_modmult_fast(t1, t1, Z); |
| |
| /* y1 * z^3 */ |
| |
| vli_modmult_fast(Y1, Y1, t1); |
| } |
| |
| /* P = (x1, y1) => 2P, (x2, y2) => P' */ |
| |
| static void xycz_initial_double(FAR uint64_t *X1, FAR uint64_t *Y1, |
| FAR uint64_t *X2, FAR uint64_t *Y2, |
| FAR uint64_t *initialz) |
| { |
| uint64_t z[NUM_ECC_DIGITS]; |
| |
| vli_set(X2, X1); |
| vli_set(Y2, Y1); |
| |
| vli_clear(z); |
| z[0] = 1; |
| if (initialz) |
| { |
| vli_set(z, initialz); |
| } |
| |
| apply_z(X1, Y1, z); |
| |
| eccpoint_double_jacobian(X1, Y1, z); |
| |
| apply_z(X2, Y2, z); |
| } |
| |
| /* Input P = (x1, y1, Z), Q = (x2, y2, Z) |
| * Output P' = (x1', y1', Z3), P + Q = (x3, y3, Z3) |
| * or P => P', Q => P + Q |
| */ |
| |
| static void xycz_add(FAR uint64_t *X1, FAR uint64_t *Y1, |
| FAR uint64_t *X2, FAR uint64_t *Y2) |
| { |
| /* t1 = X1, t2 = Y1, t3 = X2, t4 = Y2 */ |
| |
| uint64_t t5[NUM_ECC_DIGITS]; |
| |
| /* t5 = x2 - x1 */ |
| |
| vli_modsub(t5, X2, X1, g_curve_p); |
| |
| /* t5 = (x2 - x1)^2 = A */ |
| |
| vli_modsquare_fast(t5, t5); |
| |
| /* t1 = x1*A = B */ |
| |
| vli_modmult_fast(X1, X1, t5); |
| |
| /* t3 = x2*A = C */ |
| |
| vli_modmult_fast(X2, X2, t5); |
| |
| /* t4 = y2 - y1 */ |
| |
| vli_modsub(Y2, Y2, Y1, g_curve_p); |
| |
| /* t5 = (y2 - y1)^2 = D */ |
| |
| vli_modsquare_fast(t5, Y2); |
| |
| /* t5 = D - B */ |
| |
| vli_modsub(t5, t5, X1, g_curve_p); |
| |
| /* t5 = D - B - C = x3 */ |
| |
| vli_modsub(t5, t5, X2, g_curve_p); |
| |
| /* t3 = C - B */ |
| |
| vli_modsub(X2, X2, X1, g_curve_p); |
| |
| /* t2 = y1*(C - B) */ |
| |
| vli_modmult_fast(Y1, Y1, X2); |
| |
| /* t3 = B - x3 */ |
| |
| vli_modsub(X2, X1, t5, g_curve_p); |
| |
| /* t4 = (y2 - y1)*(B - x3) */ |
| |
| vli_modmult_fast(Y2, Y2, X2); |
| |
| /* t4 = y3 */ |
| |
| vli_modsub(Y2, Y2, Y1, g_curve_p); |
| vli_set(X2, t5); |
| } |
| |
| /* Input P = (x1, y1, Z), Q = (x2, y2, Z) |
| * Output P + Q = (x3, y3, Z3), P - Q = (x3', y3', Z3) |
| * or P => P - Q, Q => P + Q |
| */ |
| |
| static void xycz_addc(FAR uint64_t *X1, FAR uint64_t *Y1, |
| FAR uint64_t *X2, FAR uint64_t *Y2) |
| { |
| /* t1 = X1, t2 = Y1, t3 = X2, t4 = Y2 */ |
| |
| uint64_t t5[NUM_ECC_DIGITS]; |
| uint64_t t6[NUM_ECC_DIGITS]; |
| uint64_t t7[NUM_ECC_DIGITS]; |
| |
| /* t5 = x2 - x1 */ |
| |
| vli_modsub(t5, X2, X1, g_curve_p); |
| |
| /* t5 = (x2 - x1)^2 = A */ |
| |
| vli_modsquare_fast(t5, t5); |
| |
| /* t1 = x1*A = B */ |
| |
| vli_modmult_fast(X1, X1, t5); |
| |
| /* t3 = x2*A = C */ |
| |
| vli_modmult_fast(X2, X2, t5); |
| |
| /* t4 = y2 + y1 */ |
| |
| vli_modadd(t5, Y2, Y1, g_curve_p); |
| |
| /* t4 = y2 - y1 */ |
| |
| vli_modsub(Y2, Y2, Y1, g_curve_p); |
| |
| /* t6 = C - B */ |
| |
| vli_modsub(t6, X2, X1, g_curve_p); |
| |
| /* t2 = y1 * (C - B) */ |
| |
| vli_modmult_fast(Y1, Y1, t6); |
| |
| /* t6 = B + C */ |
| |
| vli_modadd(t6, X1, X2, g_curve_p); |
| |
| /* t3 = (y2 - y1)^2 */ |
| |
| vli_modsquare_fast(X2, Y2); |
| |
| /* t3 = x3 */ |
| |
| vli_modsub(X2, X2, t6, g_curve_p); |
| |
| /* t7 = B - x3 */ |
| |
| vli_modsub(t7, X1, X2, g_curve_p); |
| |
| /* t4 = (y2 - y1)*(B - x3) */ |
| |
| vli_modmult_fast(Y2, Y2, t7); |
| |
| /* t4 = y3 */ |
| |
| vli_modsub(Y2, Y2, Y1, g_curve_p); |
| |
| /* t7 = (y2 + y1)^2 = F */ |
| |
| vli_modsquare_fast(t7, t5); |
| |
| /* t7 = x3' */ |
| |
| vli_modsub(t7, t7, t6, g_curve_p); |
| |
| /* t6 = x3' - B */ |
| |
| vli_modsub(t6, t7, X1, g_curve_p); |
| |
| /* t6 = (y2 + y1)*(x3' - B) */ |
| |
| vli_modmult_fast(t6, t6, t5); |
| |
| /* t2 = y3' */ |
| |
| vli_modsub(Y1, t6, Y1, g_curve_p); |
| |
| vli_set(X1, t7); |
| } |
| |
| static void eccpoint_mult(FAR eccpoint_t *result, FAR eccpoint_t *point, |
| FAR uint64_t *scalar, FAR uint64_t *initialz) |
| { |
| /* R0 and R1 */ |
| |
| uint64_t rx[2][NUM_ECC_DIGITS]; |
| uint64_t ry[2][NUM_ECC_DIGITS]; |
| uint64_t z[NUM_ECC_DIGITS]; |
| int nb; |
| int i; |
| |
| vli_set(rx[1], point->x); |
| vli_set(ry[1], point->y); |
| |
| xycz_initial_double(rx[1], ry[1], rx[0], ry[0], initialz); |
| |
| for (i = vli_numbits(scalar) - 2; i > 0; --i) |
| { |
| nb = !vli_testbit(scalar, i); |
| xycz_addc(rx[1 - nb], ry[1 - nb], rx[nb], ry[nb]); |
| xycz_add(rx[nb], ry[nb], rx[1 - nb], ry[1 - nb]); |
| } |
| |
| nb = !vli_testbit(scalar, 0); |
| xycz_addc(rx[1 - nb], ry[1 - nb], rx[nb], ry[nb]); |
| |
| /* Find final 1/Z value. */ |
| |
| /* X1 - X0 */ |
| |
| vli_modsub(z, rx[1], rx[0], g_curve_p); |
| |
| /* Yb * (X1 - X0) */ |
| |
| vli_modmult_fast(z, z, ry[1 - nb]); |
| |
| /* xP * Yb * (X1 - X0) */ |
| |
| vli_modmult_fast(z, z, point->x); |
| |
| /* 1 / (xP * Yb * (X1 - X0)) */ |
| |
| vli_modinv(z, z, g_curve_p); |
| |
| /* yP / (xP * Yb * (X1 - X0)) */ |
| |
| vli_modmult_fast(z, z, point->y); |
| |
| /* Xb * yP / (xP * Yb * (X1 - X0)) */ |
| |
| vli_modmult_fast(z, z, rx[1 - nb]); |
| |
| /* End 1/Z calculation */ |
| |
| xycz_add(rx[nb], ry[nb], rx[1 - nb], ry[1 - nb]); |
| |
| apply_z(rx[0], ry[0], z); |
| |
| vli_set(result->x, rx[0]); |
| vli_set(result->y, ry[0]); |
| } |
| |
| static void ecc_bytes2native(uint64_t native[NUM_ECC_DIGITS], |
| const uint8_t bytes[ECC_BYTES]) |
| { |
| FAR const uint8_t *digit; |
| unsigned i; |
| |
| for (i = 0; i < NUM_ECC_DIGITS; ++i) |
| { |
| digit = bytes + 8 * (NUM_ECC_DIGITS - 1 - i); |
| native[i] = ((uint64_t)digit[0] << 56) | ((uint64_t)digit[1] << 48) | |
| ((uint64_t)digit[2] << 40) | ((uint64_t)digit[3] << 32) | |
| ((uint64_t)digit[4] << 24) | ((uint64_t)digit[5] << 16) | |
| ((uint64_t)digit[6] << 8) | (uint64_t)digit[7]; |
| } |
| } |
| |
| static void ecc_native2bytes(uint8_t bytes[ECC_BYTES], |
| const uint64_t native[NUM_ECC_DIGITS]) |
| { |
| FAR uint8_t *digit; |
| unsigned i; |
| |
| for (i = 0; i < NUM_ECC_DIGITS; ++i) |
| { |
| digit = bytes + 8 * (NUM_ECC_DIGITS - 1 - i); |
| digit[0] = native[i] >> 56; |
| digit[1] = native[i] >> 48; |
| digit[2] = native[i] >> 40; |
| digit[3] = native[i] >> 32; |
| digit[4] = native[i] >> 24; |
| digit[5] = native[i] >> 16; |
| digit[6] = native[i] >> 8; |
| digit[7] = native[i]; |
| } |
| } |
| |
| /* Compute a = sqrt(a) (mod g_curve_p). */ |
| |
| static void mod_sqrt(uint64_t a[NUM_ECC_DIGITS]) |
| { |
| unsigned i; |
| uint64_t l_result[NUM_ECC_DIGITS] = |
| { |
| 1 |
| }; |
| |
| uint64_t p1[NUM_ECC_DIGITS] = |
| { |
| 1 |
| }; |
| |
| /* Since g_curve_p == 3 (mod 4) for all supported curves, we can |
| * compute sqrt(a) = a^((g_curve_p + 1) / 4) (mod g_curve_p). |
| */ |
| |
| vli_add(p1, g_curve_p, p1); /* p1 = g_curve_p + 1 */ |
| for (i = vli_numbits(p1) - 1; i > 1; --i) |
| { |
| vli_modsquare_fast(l_result, l_result); |
| if (vli_testbit(p1, i)) |
| { |
| vli_modmult_fast(l_result, l_result, a); |
| } |
| } |
| |
| vli_set(a, l_result); |
| } |
| |
| static void |
| ecc_point_decompress(FAR eccpoint_t *point, |
| const uint8_t compressed[ECC_BYTES + 1]) |
| { |
| /* -a = 3 */ |
| |
| uint64_t _3[NUM_ECC_DIGITS] = |
| { |
| 3 |
| }; |
| |
| ecc_bytes2native(point->x, compressed + 1); |
| |
| /* y = x^2 */ |
| |
| vli_modsquare_fast(point->y, point->x); |
| |
| /* y = x^2 - 3 */ |
| |
| vli_modsub(point->y, point->y, _3, g_curve_p); |
| |
| /* y = x^3 - 3x */ |
| |
| vli_modmult_fast(point->y, point->y, point->x); |
| |
| /* y = x^3 - 3x + b */ |
| |
| vli_modadd(point->y, point->y, g_curve_b, g_curve_p); |
| |
| mod_sqrt(point->y); |
| |
| if ((point->y[0] & 0x01) != (compressed[0] & 0x01)) |
| { |
| vli_sub(point->y, g_curve_p, point->y); |
| } |
| } |
| |
| /* -------- ECDSA code -------- */ |
| |
| /* Computes result = (left * right) % mod. */ |
| |
| static void vli_modmult(FAR uint64_t *result, FAR uint64_t *left, |
| FAR uint64_t *right, FAR uint64_t *mod) |
| { |
| uint64_t l_product[2 * NUM_ECC_DIGITS]; |
| uint64_t l_modmultiple[2 * NUM_ECC_DIGITS]; |
| uint64_t l_carry; |
| uint l_modbits = vli_numbits(mod); |
| uint l_productbits; |
| uint l_digitshift; |
| uint l_bitshift; |
| int l_cmp; |
| |
| vli_mult(l_product, left, right); |
| l_productbits = vli_numbits(l_product + NUM_ECC_DIGITS); |
| if (l_productbits) |
| { |
| l_productbits += NUM_ECC_DIGITS * 64; |
| } |
| else |
| { |
| l_productbits = vli_numbits(l_product); |
| } |
| |
| if (l_productbits < l_modbits) |
| { |
| /* l_product < mod. */ |
| |
| vli_set(result, l_product); |
| return; |
| } |
| |
| /* Shift mod by (l_leftBits - l_modbits). |
| * This multiplies mod by the largest power of two possible |
| * while still resulting in a number less than left. |
| */ |
| |
| vli_clear(l_modmultiple); |
| vli_clear(l_modmultiple + NUM_ECC_DIGITS); |
| l_digitshift = (l_productbits - l_modbits) / 64; |
| l_bitshift = (l_productbits - l_modbits) % 64; |
| if (l_bitshift) |
| { |
| l_modmultiple[l_digitshift + NUM_ECC_DIGITS] = |
| vli_lshift(l_modmultiple + l_digitshift, mod, l_bitshift); |
| } |
| else |
| { |
| vli_set(l_modmultiple + l_digitshift, mod); |
| } |
| |
| /* Subtract all multiples of mod to get the remainder. */ |
| |
| vli_clear(result); |
| |
| /* Use result as a temp var to store 1 (for subtraction) */ |
| |
| result[0] = 1; |
| while (l_productbits > NUM_ECC_DIGITS * 64 || |
| vli_cmp(l_modmultiple, mod) >= 0) |
| { |
| l_cmp = vli_cmp(l_modmultiple + NUM_ECC_DIGITS, |
| l_product + NUM_ECC_DIGITS); |
| if (l_cmp < 0 || |
| (l_cmp == 0 && vli_cmp(l_modmultiple, l_product) <= 0)) |
| { |
| if (vli_sub(l_product, l_product, l_modmultiple)) |
| { |
| /* borrow */ |
| |
| vli_sub(l_product + NUM_ECC_DIGITS, |
| l_product + NUM_ECC_DIGITS, result); |
| } |
| |
| vli_sub(l_product + NUM_ECC_DIGITS, l_product + NUM_ECC_DIGITS, |
| l_modmultiple + NUM_ECC_DIGITS); |
| } |
| |
| l_carry = (l_modmultiple[NUM_ECC_DIGITS] & 0x01) << 63; |
| vli_rshift1(l_modmultiple + NUM_ECC_DIGITS); |
| vli_rshift1(l_modmultiple); |
| l_modmultiple[NUM_ECC_DIGITS - 1] |= l_carry; |
| --l_productbits; |
| } |
| |
| vli_set(result, l_product); |
| } |
| |
| static uint umax(uint a, uint b) |
| { |
| return a > b ? a : b; |
| } |
| |
| /**************************************************************************** |
| * Public Functions |
| ****************************************************************************/ |
| |
| int ecc_make_key(uint8_t publickey[ECC_BYTES + 1], |
| uint8_t privatekey[ECC_BYTES]) |
| { |
| uint64_t l_private[NUM_ECC_DIGITS]; |
| eccpoint_t l_public; |
| unsigned l_tries = 0; |
| |
| memset(&l_public, 0, sizeof(eccpoint_t)); |
| |
| do |
| { |
| if (l_tries++ >= MAX_TRIES) |
| { |
| return 0; |
| } |
| |
| arc4random_buf(l_private, NUM_ECC_DIGITS); |
| |
| if (vli_iszero(l_private)) |
| { |
| continue; |
| } |
| |
| /* Make sure the private key is in the range [1, n-1]. |
| * For the supported curves, n is always large enough that we only |
| * need to subtract once at most. |
| */ |
| |
| if (vli_cmp(g_curve_n, l_private) != 1) |
| { |
| vli_sub(l_private, l_private, g_curve_n); |
| } |
| |
| eccpoint_mult(&l_public, &g_curve_g, l_private, NULL); |
| } |
| while (eccpoint_iszero(&l_public)); |
| |
| ecc_native2bytes(privatekey, l_private); |
| ecc_native2bytes(publickey + 1, l_public.x); |
| publickey[0] = 2 + (l_public.y[0] & 0x01); |
| return 1; |
| } |
| |
| int ecc_make_key_uncomp(uint8_t publickey_x[ECC_BYTES], |
| uint8_t publickey_y[ECC_BYTES], |
| uint8_t privatekey[ECC_BYTES]) |
| { |
| uint64_t l_private[NUM_ECC_DIGITS]; |
| eccpoint_t l_public; |
| unsigned l_tries = 0; |
| |
| do |
| { |
| if (l_tries++ >= MAX_TRIES) |
| { |
| return 0; |
| } |
| |
| arc4random_buf(l_private, NUM_ECC_DIGITS); |
| |
| if (vli_iszero(l_private)) |
| { |
| continue; |
| } |
| |
| /* Make sure the private key is in the range [1, n-1]. |
| * For the supported curves, n is always large enough that we only |
| * need to subtract once at most. |
| */ |
| |
| if (vli_cmp(g_curve_n, l_private) != 1) |
| { |
| vli_sub(l_private, l_private, g_curve_n); |
| } |
| |
| eccpoint_mult(&l_public, &g_curve_g, l_private, NULL); |
| } |
| while (eccpoint_iszero(&l_public)); |
| |
| ecc_native2bytes(privatekey, l_private); |
| ecc_native2bytes(publickey_x, l_public.x); |
| ecc_native2bytes(publickey_y, l_public.y); |
| return 1; |
| } |
| |
| int ecdh_shared_secret(const uint8_t publickey[ECC_BYTES + 1], |
| const uint8_t privatekey[ECC_BYTES], |
| uint8_t secret[ECC_BYTES]) |
| { |
| eccpoint_t l_product; |
| eccpoint_t l_public; |
| uint64_t l_private[NUM_ECC_DIGITS]; |
| uint64_t l_random[NUM_ECC_DIGITS]; |
| |
| memset(&l_product, 0, sizeof(eccpoint_t)); |
| memset(&l_public, 0, sizeof(eccpoint_t)); |
| |
| arc4random_buf(l_random, NUM_ECC_DIGITS); |
| ecc_point_decompress(&l_public, publickey); |
| ecc_bytes2native(l_private, privatekey); |
| |
| eccpoint_mult(&l_product, &l_public, l_private, l_random); |
| |
| ecc_native2bytes(secret, l_product.x); |
| |
| return !eccpoint_iszero(&l_product); |
| } |
| |
| int ecdsa_sign(const uint8_t privatekey[ECC_BYTES], |
| const uint8_t hash[ECC_BYTES], |
| uint8_t signature[ECC_BYTES * 2]) |
| { |
| uint64_t k[NUM_ECC_DIGITS]; |
| uint64_t l_tmp[NUM_ECC_DIGITS]; |
| uint64_t l_s[NUM_ECC_DIGITS]; |
| unsigned l_tries = 0; |
| eccpoint_t p; |
| |
| memset(&p, 0, sizeof(eccpoint_t)); |
| |
| do |
| { |
| if (l_tries++ >= MAX_TRIES) |
| { |
| return 0; |
| } |
| |
| arc4random_buf(k, NUM_ECC_DIGITS); |
| |
| if (vli_iszero(k)) |
| { |
| continue; |
| } |
| |
| if (vli_cmp(g_curve_n, k) != 1) |
| { |
| vli_sub(k, k, g_curve_n); |
| } |
| |
| /* tmp = k * G */ |
| |
| eccpoint_mult(&p, &g_curve_g, k, NULL); |
| |
| /* r = x1 (mod n) */ |
| |
| if (vli_cmp(g_curve_n, p.x) != 1) |
| { |
| vli_sub(p.x, p.x, g_curve_n); |
| } |
| } |
| while (vli_iszero(p.x)); |
| |
| ecc_native2bytes(signature, p.x); |
| |
| ecc_bytes2native(l_tmp, privatekey); |
| vli_modmult(l_s, p.x, l_tmp, g_curve_n); /* s = r*d */ |
| ecc_bytes2native(l_tmp, hash); |
| vli_modadd(l_s, l_tmp, l_s, g_curve_n); /* s = e + r*d */ |
| |
| /* k = 1 / k */ |
| |
| vli_modinv(k, k, g_curve_n); |
| |
| /* s = (e + r*d) / k */ |
| |
| vli_modmult(l_s, l_s, k, g_curve_n); |
| ecc_native2bytes(signature + ECC_BYTES, l_s); |
| |
| return 1; |
| } |
| |
| int ecdsa_verify(const uint8_t publickey[ECC_BYTES + 1], |
| const uint8_t hash[ECC_BYTES], |
| const uint8_t signature[ECC_BYTES * 2]) |
| { |
| uint64_t u1[NUM_ECC_DIGITS]; |
| uint64_t u2[NUM_ECC_DIGITS]; |
| uint64_t z[NUM_ECC_DIGITS]; |
| uint64_t rx[NUM_ECC_DIGITS]; |
| uint64_t ry[NUM_ECC_DIGITS]; |
| uint64_t tx[NUM_ECC_DIGITS]; |
| uint64_t ty[NUM_ECC_DIGITS]; |
| uint64_t tz[NUM_ECC_DIGITS]; |
| uint64_t l_r[NUM_ECC_DIGITS]; |
| uint64_t l_s[NUM_ECC_DIGITS]; |
| uint l_numbits; |
| eccpoint_t *l_point; |
| eccpoint_t l_public; |
| eccpoint_t l_sum; |
| int l_index; |
| int i; |
| |
| /* Use Shamir's trick to calculate u1*G + u2*Q */ |
| |
| eccpoint_t *l_points[4] = |
| { |
| NULL, &g_curve_g, &l_public, &l_sum |
| }; |
| |
| ecc_point_decompress(&l_public, publickey); |
| ecc_bytes2native(l_r, signature); |
| ecc_bytes2native(l_s, signature + ECC_BYTES); |
| |
| if (vli_iszero(l_r) || vli_iszero(l_s)) |
| { |
| /* r, s must not be 0. */ |
| |
| return 0; |
| } |
| |
| if (vli_cmp(g_curve_n, l_r) != 1 || vli_cmp(g_curve_n, l_s) != 1) |
| { |
| /* r, s must be < n. */ |
| |
| return 0; |
| } |
| |
| /* Calculate u1 and u2. */ |
| |
| vli_modinv(z, l_s, g_curve_n); /* Z = s^-1 */ |
| ecc_bytes2native(u1, hash); |
| vli_modmult(u1, u1, z, g_curve_n); /* u1 = e/s */ |
| |
| /* u2 = r/s */ |
| |
| vli_modmult(u2, l_r, z, g_curve_n); |
| |
| /* Calculate l_sum = G + Q. */ |
| |
| vli_set(l_sum.x, l_public.x); |
| vli_set(l_sum.y, l_public.y); |
| vli_set(tx, g_curve_g.x); |
| vli_set(ty, g_curve_g.y); |
| vli_modsub(z, l_sum.x, tx, g_curve_p); /* Z = x2 - x1 */ |
| xycz_add(tx, ty, l_sum.x, l_sum.y); |
| vli_modinv(z, z, g_curve_p); /* Z = 1/Z */ |
| apply_z(l_sum.x, l_sum.y, z); |
| |
| l_numbits = umax(vli_numbits(u1), vli_numbits(u2)); |
| |
| l_point = l_points[(!!vli_testbit(u1, l_numbits - 1)) | |
| ((!!vli_testbit(u2, l_numbits - 1)) << 1)]; |
| vli_set(rx, l_point->x); |
| vli_set(ry, l_point->y); |
| vli_clear(z); |
| z[0] = 1; |
| |
| for (i = l_numbits - 2; i >= 0; --i) |
| { |
| eccpoint_double_jacobian(rx, ry, z); |
| |
| l_index = (!!vli_testbit(u1, i)) | ((!!vli_testbit(u2, i)) << 1); |
| l_point = l_points[l_index]; |
| if (l_point) |
| { |
| vli_set(tx, l_point->x); |
| vli_set(ty, l_point->y); |
| apply_z(tx, ty, z); |
| vli_modsub(tz, rx, tx, g_curve_p); /* Z = x2 - x1 */ |
| xycz_add(tx, ty, rx, ry); |
| vli_modmult_fast(z, z, tz); |
| } |
| } |
| |
| vli_modinv(z, z, g_curve_p); /* Z = 1/Z */ |
| apply_z(rx, ry, z); |
| |
| /* v = x1 (mod n) */ |
| |
| if (vli_cmp(g_curve_n, rx) != 1) |
| { |
| vli_sub(rx, rx, g_curve_n); |
| } |
| |
| /* Accept only if v == r. */ |
| |
| return vli_cmp(rx, l_r) == 0; |
| } |