mirror of
https://github.com/apache/nuttx.git
synced 2026-05-12 01:42:49 +08:00
85ba80a90e
Add BSD-2-Clause SPDX license identifier to ECC source and header files. Signed-off-by: makejian <makejian@xiaomi.com>
1781 lines
41 KiB
C
1781 lines
41 KiB
C
/****************************************************************************
|
|
* crypto/ecc.c
|
|
*
|
|
* SPDX-License-Identifier: BSD-2-Clause
|
|
*
|
|
* Copyright (c) 2013, Kenneth MacKay All rights reserved.
|
|
*
|
|
* Redistribution and use in source and binary forms, with or without
|
|
* modification, are permitted provided that the following conditions are
|
|
* met: Redistributions of source code must retain the above copyright
|
|
* notice, this list of conditions and the following disclaimer.
|
|
* Redistributions in binary form must reproduce the above copyright notice,
|
|
* this list of conditions and the following disclaimer in the documentation
|
|
* and/or other materials provided with the distribution.
|
|
|
|
* THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
|
|
* "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
|
|
* LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
|
|
* A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
|
|
* SPECIAL, HOLDER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT,
|
|
* INCIDENTAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
|
|
* LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
|
|
* DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
|
|
* THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
|
|
* (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
|
|
* OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
|
****************************************************************************/
|
|
|
|
/****************************************************************************
|
|
* Included Files
|
|
****************************************************************************/
|
|
|
|
#include <fcntl.h>
|
|
#include <stdlib.h>
|
|
#include <string.h>
|
|
#include <unistd.h>
|
|
#include <sys/types.h>
|
|
|
|
#include <crypto/ecc.h>
|
|
#include <nuttx/macro.h>
|
|
|
|
/****************************************************************************
|
|
* Pre-processor Definitions
|
|
****************************************************************************/
|
|
|
|
#define NUM_ECC_DIGITS (ECC_BYTES / 8)
|
|
#define MAX_TRIES 16
|
|
|
|
#define EVEN(vli) (!(vli[0] & 1))
|
|
|
|
#define curve_p_16 { 0xFFFFFFFFFFFFFFFF, 0xFFFFFFFDFFFFFFFF }
|
|
#define curve_p_24 { 0xFFFFFFFFFFFFFFFFull, 0xFFFFFFFFFFFFFFFEull, \
|
|
0xFFFFFFFFFFFFFFFFull }
|
|
#define curve_p_32 { 0xFFFFFFFFFFFFFFFFull, 0x00000000FFFFFFFFull, \
|
|
0x0000000000000000ull, 0xFFFFFFFF00000001ull }
|
|
#define curve_p_48 { 0x00000000FFFFFFFF, 0xFFFFFFFF00000000, \
|
|
0xFFFFFFFFFFFFFFFE, 0xFFFFFFFFFFFFFFFF, \
|
|
0xFFFFFFFFFFFFFFFF, 0xFFFFFFFFFFFFFFFF }
|
|
|
|
#define curve_b_16 { 0xD824993C2CEE5ED3, 0xE87579C11079F43D }
|
|
#define curve_b_24 { 0xFEB8DEECC146B9B1ull, 0x0FA7E9AB72243049ull, \
|
|
0x64210519E59C80E7ull }
|
|
#define curve_b_32 { 0x3BCE3C3E27D2604Bull, 0x651D06B0CC53B0F6ull, \
|
|
0xB3EBBD55769886BCull, 0x5AC635D8AA3A93E7ull }
|
|
#define curve_b_48 { 0x2A85C8EDD3EC2AEF, 0xC656398D8A2ED19D, \
|
|
0x0314088F5013875A, 0x181D9C6EFE814112, \
|
|
0x988E056BE3F82D19, 0xB3312FA7E23EE7E4 }
|
|
|
|
#define curve_g_16 { \
|
|
{ 0x0C28607CA52C5B86, 0x161FF7528B899B2D }, \
|
|
{ 0xC02DA292DDED7A83, 0xCF5AC8395BAFEB13 }}
|
|
|
|
#define curve_g_24 { \
|
|
{ 0xF4FF0AFD82FF1012ull, 0x7CBF20EB43A18800ull, 0x188DA80EB03090F6ull }, \
|
|
{ 0x73F977A11E794811ull, 0x631011ED6B24CDD5ull, 0x07192B95FFC8DA78ull }}
|
|
|
|
#define curve_g_32 { \
|
|
{ 0xF4A13945D898C296ull, 0x77037D812DEB33A0ull, \
|
|
0xF8BCE6E563A440F2ull, 0x6B17D1F2E12C4247ull }, \
|
|
{ 0xCBB6406837BF51F5ull, 0x2BCE33576B315ECEull, \
|
|
0x8EE7EB4A7C0F9E16ull, 0x4FE342E2FE1A7F9Bull }}
|
|
|
|
#define curve_g_48 { \
|
|
{ 0x3A545E3872760AB7, 0x5502F25DBF55296C, 0x59F741E082542A38, \
|
|
0x6E1D3B628BA79B98, 0x8EB1C71EF320AD74, 0xAA87CA22BE8B0537}, \
|
|
{ 0x7A431D7C90EA0E5F, 0x0A60B1CE1D7E819D, 0xE9DA3113B5F0B8C0, \
|
|
0xF8F41DBD289A147C, 0x5D9E98BF9292DC29, 0x3617DE4A96262C6F }}
|
|
|
|
#define curve_n_16 { 0x75A30D1B9038A115, 0xFFFFFFFE00000000 }
|
|
#define curve_n_24 { 0x146BC9B1B4D22831ull, 0xFFFFFFFF99DEF836ull, \
|
|
0xFFFFFFFFFFFFFFFFull }
|
|
#define curve_n_32 { 0xF3B9CAC2FC632551ull, 0xBCE6FAADA7179E84ull, \
|
|
0xFFFFFFFFFFFFFFFFull, 0xFFFFFFFF00000000ull }
|
|
#define curve_n_48 { 0xECEC196ACCC52973, 0x581A0DB248B0A77A, \
|
|
0xC7634D81F4372DDF, 0xFFFFFFFFFFFFFFFF, \
|
|
0xFFFFFFFFFFFFFFFF, 0xFFFFFFFFFFFFFFFF }
|
|
|
|
#if defined(__SIZEOF_INT128__)
|
|
# define SUPPORTS_INT128 1
|
|
#else
|
|
# define SUPPORTS_INT128 0
|
|
#endif
|
|
|
|
/****************************************************************************
|
|
* Private Type Definitions
|
|
****************************************************************************/
|
|
|
|
#if SUPPORTS_INT128
|
|
typedef unsigned __int128 uint128_t;
|
|
#else
|
|
typedef struct
|
|
{
|
|
uint64_t m_low;
|
|
uint64_t m_high;
|
|
} uint128_t;
|
|
#endif
|
|
|
|
typedef struct
|
|
{
|
|
uint64_t x[NUM_ECC_DIGITS];
|
|
uint64_t y[NUM_ECC_DIGITS];
|
|
} eccpoint_t;
|
|
|
|
/****************************************************************************
|
|
* Private Data
|
|
****************************************************************************/
|
|
|
|
static uint64_t g_curve_p[NUM_ECC_DIGITS] = CONCATENATE(curve_p_, ECC_CURVE);
|
|
static uint64_t g_curve_b[NUM_ECC_DIGITS] = CONCATENATE(curve_b_, ECC_CURVE);
|
|
static uint64_t g_curve_n[NUM_ECC_DIGITS] = CONCATENATE(curve_n_, ECC_CURVE);
|
|
static eccpoint_t g_curve_g = CONCATENATE(curve_g_, ECC_CURVE);
|
|
|
|
/****************************************************************************
|
|
* Private Functions
|
|
****************************************************************************/
|
|
|
|
static void vli_clear(FAR uint64_t *vli)
|
|
{
|
|
uint i;
|
|
|
|
for (i = 0; i < NUM_ECC_DIGITS; ++i)
|
|
{
|
|
vli[i] = 0;
|
|
}
|
|
}
|
|
|
|
/* Returns 1 if vli == 0, 0 otherwise. */
|
|
|
|
static int vli_iszero(FAR uint64_t *vli)
|
|
{
|
|
uint i;
|
|
|
|
for (i = 0; i < NUM_ECC_DIGITS; ++i)
|
|
{
|
|
if (vli[i])
|
|
{
|
|
return 0;
|
|
}
|
|
}
|
|
|
|
return 1;
|
|
}
|
|
|
|
/* Returns nonzero if bit bit of vli is set. */
|
|
|
|
static uint64_t vli_testbit(FAR uint64_t *vli, uint bit)
|
|
{
|
|
return vli[bit / 64] & ((uint64_t)1 << (bit % 64));
|
|
}
|
|
|
|
/* Counts the number of 64-bit "digits" in vli. */
|
|
|
|
static uint vli_numdigits(FAR uint64_t *vli)
|
|
{
|
|
int i;
|
|
|
|
/* Search from the end until we find a non-zero digit.
|
|
* We do it in reverse because we expect that most digits
|
|
* will be nonzero.
|
|
*/
|
|
|
|
for (i = NUM_ECC_DIGITS - 1; i >= 0 && vli[i] == 0; --i)
|
|
{
|
|
}
|
|
|
|
return i + 1;
|
|
}
|
|
|
|
/* Counts the number of bits required for vli. */
|
|
|
|
static uint vli_numbits(FAR uint64_t *vli)
|
|
{
|
|
uint64_t l_digit;
|
|
uint l_numdigits = vli_numdigits(vli);
|
|
uint i;
|
|
|
|
if (l_numdigits == 0)
|
|
{
|
|
return 0;
|
|
}
|
|
|
|
l_digit = vli[l_numdigits - 1];
|
|
for (i = 0; l_digit; ++i)
|
|
{
|
|
l_digit >>= 1;
|
|
}
|
|
|
|
return (l_numdigits - 1) * 64 + i;
|
|
}
|
|
|
|
/* Sets dest = src. */
|
|
|
|
static void vli_set(FAR uint64_t *dest, FAR uint64_t *src)
|
|
{
|
|
uint i;
|
|
|
|
for (i = 0; i < NUM_ECC_DIGITS; ++i)
|
|
{
|
|
dest[i] = src[i];
|
|
}
|
|
}
|
|
|
|
/* Returns sign of left - right. */
|
|
|
|
static int vli_cmp(FAR uint64_t *left, FAR uint64_t *right)
|
|
{
|
|
int i;
|
|
|
|
for (i = NUM_ECC_DIGITS - 1; i >= 0; --i)
|
|
{
|
|
if (left[i] > right[i])
|
|
{
|
|
return 1;
|
|
}
|
|
else if (left[i] < right[i])
|
|
{
|
|
return -1;
|
|
}
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
/* Computes result = in << c, returning carry.
|
|
* Can modify in place (if result == in). 0 < shift < 64.
|
|
*/
|
|
|
|
static uint64_t vli_lshift(FAR uint64_t *result, FAR uint64_t *in,
|
|
uint shift)
|
|
{
|
|
uint64_t l_carry = 0;
|
|
uint64_t l_temp;
|
|
uint i;
|
|
|
|
for (i = 0; i < NUM_ECC_DIGITS; ++i)
|
|
{
|
|
l_temp = in[i];
|
|
result[i] = (l_temp << shift) | l_carry;
|
|
l_carry = l_temp >> (64 - shift);
|
|
}
|
|
|
|
return l_carry;
|
|
}
|
|
|
|
/* Computes vli = vli >> 1. */
|
|
|
|
static void vli_rshift1(FAR uint64_t *vli)
|
|
{
|
|
FAR uint64_t *l_end = vli;
|
|
uint64_t l_carry = 0;
|
|
uint64_t l_temp;
|
|
|
|
vli += NUM_ECC_DIGITS;
|
|
while (vli-- > l_end)
|
|
{
|
|
l_temp = *vli;
|
|
*vli = (l_temp >> 1) | l_carry;
|
|
l_carry = l_temp << 63;
|
|
}
|
|
}
|
|
|
|
/* Computes result = left + right, returning carry. Can modify in place. */
|
|
|
|
static uint64_t vli_add(FAR uint64_t *result, FAR uint64_t *left,
|
|
FAR uint64_t *right)
|
|
{
|
|
uint64_t l_carry = 0;
|
|
uint64_t l_sum;
|
|
uint i;
|
|
|
|
for (i = 0; i < NUM_ECC_DIGITS; ++i)
|
|
{
|
|
l_sum = left[i] + right[i] + l_carry;
|
|
if (l_sum != left[i])
|
|
{
|
|
l_carry = (l_sum < left[i]);
|
|
}
|
|
|
|
result[i] = l_sum;
|
|
}
|
|
|
|
return l_carry;
|
|
}
|
|
|
|
/* Computes result = left - right, returning borrow. Can modify in place. */
|
|
|
|
static uint64_t vli_sub(FAR uint64_t *result, FAR uint64_t *left,
|
|
FAR uint64_t *right)
|
|
{
|
|
uint64_t l_borrow = 0;
|
|
uint64_t l_diff;
|
|
uint i;
|
|
|
|
for (i = 0; i < NUM_ECC_DIGITS; ++i)
|
|
{
|
|
l_diff = left[i] - right[i] - l_borrow;
|
|
if (l_diff != left[i])
|
|
{
|
|
l_borrow = (l_diff > left[i]);
|
|
}
|
|
|
|
result[i] = l_diff;
|
|
}
|
|
|
|
return l_borrow;
|
|
}
|
|
|
|
#if SUPPORTS_INT128
|
|
|
|
/* Computes result = left * right. */
|
|
|
|
static void vli_mult(FAR uint64_t *result, FAR uint64_t *left,
|
|
FAR uint64_t *right)
|
|
{
|
|
uint128_t l_product;
|
|
uint128_t r01 = 0;
|
|
uint64_t r2 = 0;
|
|
uint l_min;
|
|
uint i;
|
|
uint k;
|
|
|
|
/* Compute each digit of result in sequence, maintaining the carries. */
|
|
|
|
for (k = 0; k < NUM_ECC_DIGITS * 2 - 1; ++k)
|
|
{
|
|
l_min = (k < NUM_ECC_DIGITS ? 0 : (k + 1) - NUM_ECC_DIGITS);
|
|
for (i = l_min; i <= k && i < NUM_ECC_DIGITS; ++i)
|
|
{
|
|
l_product = (uint128_t)left[i] * right[k - i];
|
|
r01 += l_product;
|
|
r2 += (r01 < l_product);
|
|
}
|
|
|
|
result[k] = (uint64_t)r01;
|
|
r01 = (r01 >> 64) | (((uint128_t)r2) << 64);
|
|
r2 = 0;
|
|
}
|
|
|
|
result[NUM_ECC_DIGITS * 2 - 1] = (uint64_t)r01;
|
|
}
|
|
|
|
/* Computes result = left^2. */
|
|
|
|
static void vli_square(FAR uint64_t *result, FAR uint64_t *left)
|
|
{
|
|
uint128_t l_product;
|
|
uint128_t r01 = 0;
|
|
uint64_t r2 = 0;
|
|
uint l_min;
|
|
uint i;
|
|
uint k;
|
|
|
|
for (k = 0; k < NUM_ECC_DIGITS * 2 - 1; ++k)
|
|
{
|
|
l_min = (k < NUM_ECC_DIGITS ? 0 : (k + 1) - NUM_ECC_DIGITS);
|
|
for (i = l_min; i <= k && i <= k - i; ++i)
|
|
{
|
|
l_product = (uint128_t)left[i] * left[k - i];
|
|
if (i < k - i)
|
|
{
|
|
r2 += l_product >> 127;
|
|
l_product *= 2;
|
|
}
|
|
|
|
r01 += l_product;
|
|
r2 += (r01 < l_product);
|
|
}
|
|
|
|
result[k] = (uint64_t)r01;
|
|
r01 = (r01 >> 64) | (((uint128_t)r2) << 64);
|
|
r2 = 0;
|
|
}
|
|
|
|
result[NUM_ECC_DIGITS * 2 - 1] = (uint64_t)r01;
|
|
}
|
|
|
|
#else /* #if SUPPORTS_INT128 */
|
|
|
|
static uint128_t mul_64_64(uint64_t left, uint64_t right)
|
|
{
|
|
uint128_t l_result;
|
|
uint64_t a0 = left & 0xffffffffull;
|
|
uint64_t a1 = left >> 32;
|
|
uint64_t b0 = right & 0xffffffffull;
|
|
uint64_t b1 = right >> 32;
|
|
uint64_t m0 = a0 * b0;
|
|
uint64_t m1 = a0 * b1;
|
|
uint64_t m2 = a1 * b0;
|
|
uint64_t m3 = a1 * b1;
|
|
|
|
m2 += (m0 >> 32);
|
|
m2 += m1;
|
|
if (m2 < m1)
|
|
{
|
|
m3 += 0x100000000ull;
|
|
}
|
|
|
|
l_result.m_low = (m0 & 0xffffffffull) | (m2 << 32);
|
|
l_result.m_high = m3 + (m2 >> 32);
|
|
|
|
return l_result;
|
|
}
|
|
|
|
static uint128_t add_128_128(uint128_t a, uint128_t b)
|
|
{
|
|
uint128_t l_result;
|
|
|
|
l_result.m_low = a.m_low + b.m_low;
|
|
l_result.m_high = a.m_high + b.m_high + (l_result.m_low < a.m_low);
|
|
return l_result;
|
|
}
|
|
|
|
static void vli_mult(FAR uint64_t *result, FAR uint64_t *left,
|
|
FAR uint64_t *right)
|
|
{
|
|
uint64_t r2 = 0;
|
|
uint i;
|
|
uint k;
|
|
uint128_t l_product;
|
|
uint128_t r01 =
|
|
{
|
|
0, 0
|
|
};
|
|
|
|
/* Compute each digit of result in sequence, maintaining the carries. */
|
|
|
|
for (k = 0; k < NUM_ECC_DIGITS * 2 - 1; ++k)
|
|
{
|
|
uint l_min = (k < NUM_ECC_DIGITS ? 0 : (k + 1) - NUM_ECC_DIGITS);
|
|
for (i = l_min; i <= k && i < NUM_ECC_DIGITS; ++i)
|
|
{
|
|
l_product = mul_64_64(left[i], right[k - i]);
|
|
r01 = add_128_128(r01, l_product);
|
|
r2 += (r01.m_high < l_product.m_high);
|
|
}
|
|
|
|
result[k] = r01.m_low;
|
|
r01.m_low = r01.m_high;
|
|
r01.m_high = r2;
|
|
r2 = 0;
|
|
}
|
|
|
|
result[NUM_ECC_DIGITS * 2 - 1] = r01.m_low;
|
|
}
|
|
|
|
static void vli_square(FAR uint64_t *result, FAR uint64_t *left)
|
|
{
|
|
uint64_t r2 = 0;
|
|
uint l_min;
|
|
uint i;
|
|
uint k;
|
|
uint128_t l_product;
|
|
uint128_t r01 =
|
|
{
|
|
0, 0
|
|
};
|
|
|
|
for (k = 0; k < NUM_ECC_DIGITS * 2 - 1; ++k)
|
|
{
|
|
l_min = (k < NUM_ECC_DIGITS ? 0 : (k + 1) - NUM_ECC_DIGITS);
|
|
for (i = l_min; i <= k && i <= k - i; ++i)
|
|
{
|
|
l_product = mul_64_64(left[i], left[k - i]);
|
|
if (i < k - i)
|
|
{
|
|
r2 += l_product.m_high >> 63;
|
|
l_product.m_high = (l_product.m_high << 1) |
|
|
(l_product.m_low >> 63);
|
|
l_product.m_low <<= 1;
|
|
}
|
|
|
|
r01 = add_128_128(r01, l_product);
|
|
r2 += (r01.m_high < l_product.m_high);
|
|
}
|
|
|
|
result[k] = r01.m_low;
|
|
r01.m_low = r01.m_high;
|
|
r01.m_high = r2;
|
|
r2 = 0;
|
|
}
|
|
|
|
result[NUM_ECC_DIGITS * 2 - 1] = r01.m_low;
|
|
}
|
|
|
|
#endif /* SUPPORTS_INT128 */
|
|
|
|
/* Computes result = (left + right) % mod.
|
|
* Assumes that left < mod and right < mod, result != mod.
|
|
*/
|
|
|
|
static void vli_modadd(FAR uint64_t *result, FAR uint64_t *left,
|
|
FAR uint64_t *right, FAR uint64_t *mod)
|
|
{
|
|
uint64_t l_carry = vli_add(result, left, right);
|
|
|
|
if (l_carry || vli_cmp(result, mod) >= 0)
|
|
{
|
|
/* result > mod (result = mod + remainder),
|
|
* so subtract mod to get remainder.
|
|
*/
|
|
|
|
vli_sub(result, result, mod);
|
|
}
|
|
}
|
|
|
|
/* Computes result = (left - right) % mod.
|
|
* Assumes that left < mod and right < mod, result != mod.
|
|
*/
|
|
|
|
static void vli_modsub(FAR uint64_t *result, FAR uint64_t *left,
|
|
FAR uint64_t *right, FAR uint64_t *mod)
|
|
{
|
|
uint64_t l_borrow = vli_sub(result, left, right);
|
|
|
|
if (l_borrow)
|
|
{
|
|
/* In this case, result == -diff == (max int) - diff.
|
|
* Since -x % d == d - x, we can get the correct result from
|
|
* result + mod (with overflow).
|
|
*/
|
|
|
|
vli_add(result, result, mod);
|
|
}
|
|
}
|
|
|
|
#if ECC_CURVE == secp128r1
|
|
|
|
/* Computes result = product % g_curve_p.
|
|
* See algorithm 5 and 6 from
|
|
* http://www.isys.uni-klu.ac.at/PDF/2001-0126-MT.pdf
|
|
*/
|
|
|
|
static void vli_mmod_fast(FAR uint64_t *result,
|
|
FAR uint64_t *product)
|
|
{
|
|
uint64_t l_tmp[NUM_ECC_DIGITS];
|
|
int l_carry;
|
|
|
|
vli_set(result, product);
|
|
|
|
l_tmp[0] = product[2];
|
|
l_tmp[1] = (product[3] & 0x1ffffffffull) | (product[2] << 33);
|
|
l_carry = vli_add(result, result, l_tmp);
|
|
|
|
l_tmp[0] = (product[2] >> 31) | (product[3] << 33);
|
|
l_tmp[1] = (product[3] >> 31) |
|
|
((product[2] & 0xffffffff80000000ull) << 2);
|
|
l_carry += vli_add(result, result, l_tmp);
|
|
|
|
l_tmp[0] = (product[2] >> 62) | (product[3] << 2);
|
|
l_tmp[1] = (product[3] >> 62) |
|
|
((product[2] & 0xc000000000000000ull) >> 29) |
|
|
(product[3] << 35);
|
|
l_carry += vli_add(result, result, l_tmp);
|
|
|
|
l_tmp[0] = (product[3] >> 29);
|
|
l_tmp[1] = ((product[3] & 0xffffffffe0000000ull) << 4);
|
|
l_carry += vli_add(result, result, l_tmp);
|
|
|
|
l_tmp[0] = (product[3] >> 60);
|
|
l_tmp[1] = (product[3] & 0xfffffffe00000000ull);
|
|
l_carry += vli_add(result, result, l_tmp);
|
|
|
|
l_tmp[0] = 0;
|
|
l_tmp[1] = ((product[3] & 0xf000000000000000ull) >> 27);
|
|
l_carry += vli_add(result, result, l_tmp);
|
|
|
|
while (l_carry || vli_cmp(g_curve_p, result) != 1)
|
|
{
|
|
l_carry -= vli_sub(result, result, g_curve_p);
|
|
}
|
|
}
|
|
|
|
#elif ECC_CURVE == secp192r1
|
|
|
|
/* Computes result = product % g_curve_p.
|
|
* See algorithm 5 and 6 from
|
|
* http://www.isys.uni-klu.ac.at/PDF/2001-0126-MT.pdf
|
|
*/
|
|
|
|
static void vli_mmod_fast(FAR uint64_t *result,
|
|
FAR uint64_t *product)
|
|
{
|
|
uint64_t l_tmp[NUM_ECC_DIGITS];
|
|
int l_carry;
|
|
|
|
vli_set(result, product);
|
|
|
|
vli_set(l_tmp, &product[3]);
|
|
l_carry = vli_add(result, result, l_tmp);
|
|
|
|
l_tmp[0] = 0;
|
|
l_tmp[1] = product[3];
|
|
l_tmp[2] = product[4];
|
|
l_carry += vli_add(result, result, l_tmp);
|
|
|
|
l_tmp[0] = l_tmp[1] = product[5];
|
|
l_tmp[2] = 0;
|
|
l_carry += vli_add(result, result, l_tmp);
|
|
|
|
while (l_carry || vli_cmp(g_curve_p, result) != 1)
|
|
{
|
|
l_carry -= vli_sub(result, result, g_curve_p);
|
|
}
|
|
}
|
|
|
|
#elif ECC_CURVE == secp256r1
|
|
|
|
/* Computes result = product % g_curve_p
|
|
* from http://www.nsa.gov/ia/_files/nist-routines.pdf
|
|
*/
|
|
|
|
static void vli_mmod_fast(FAR uint64_t *result,
|
|
FAR uint64_t *product)
|
|
{
|
|
uint64_t l_tmp[NUM_ECC_DIGITS];
|
|
int l_carry;
|
|
|
|
/* t */
|
|
|
|
vli_set(result, product);
|
|
|
|
/* s1 */
|
|
|
|
l_tmp[0] = 0;
|
|
l_tmp[1] = product[5] & 0xffffffff00000000ull;
|
|
l_tmp[2] = product[6];
|
|
l_tmp[3] = product[7];
|
|
l_carry = vli_lshift(l_tmp, l_tmp, 1);
|
|
l_carry += vli_add(result, result, l_tmp);
|
|
|
|
/* s2 */
|
|
|
|
l_tmp[1] = product[6] << 32;
|
|
l_tmp[2] = (product[6] >> 32) | (product[7] << 32);
|
|
l_tmp[3] = product[7] >> 32;
|
|
l_carry += vli_lshift(l_tmp, l_tmp, 1);
|
|
l_carry += vli_add(result, result, l_tmp);
|
|
|
|
/* s3 */
|
|
|
|
l_tmp[0] = product[4];
|
|
l_tmp[1] = product[5] & 0xffffffff;
|
|
l_tmp[2] = 0;
|
|
l_tmp[3] = product[7];
|
|
l_carry += vli_add(result, result, l_tmp);
|
|
|
|
/* s4 */
|
|
|
|
l_tmp[0] = (product[4] >> 32) | (product[5] << 32);
|
|
l_tmp[1] = (product[5] >> 32) | (product[6] & 0xffffffff00000000ull);
|
|
l_tmp[2] = product[7];
|
|
l_tmp[3] = (product[6] >> 32) | (product[4] << 32);
|
|
l_carry += vli_add(result, result, l_tmp);
|
|
|
|
/* d1 */
|
|
|
|
l_tmp[0] = (product[5] >> 32) | (product[6] << 32);
|
|
l_tmp[1] = (product[6] >> 32);
|
|
l_tmp[2] = 0;
|
|
l_tmp[3] = (product[4] & 0xffffffff) | (product[5] << 32);
|
|
l_carry -= vli_sub(result, result, l_tmp);
|
|
|
|
/* d2 */
|
|
|
|
l_tmp[0] = product[6];
|
|
l_tmp[1] = product[7];
|
|
l_tmp[2] = 0;
|
|
l_tmp[3] = (product[4] >> 32) | (product[5] & 0xffffffff00000000ull);
|
|
l_carry -= vli_sub(result, result, l_tmp);
|
|
|
|
/* d3 */
|
|
|
|
l_tmp[0] = (product[6] >> 32) | (product[7] << 32);
|
|
l_tmp[1] = (product[7] >> 32) | (product[4] << 32);
|
|
l_tmp[2] = (product[4] >> 32) | (product[5] << 32);
|
|
l_tmp[3] = (product[6] << 32);
|
|
l_carry -= vli_sub(result, result, l_tmp);
|
|
|
|
/* d4 */
|
|
|
|
l_tmp[0] = product[7];
|
|
l_tmp[1] = product[4] & 0xffffffff00000000ull;
|
|
l_tmp[2] = product[5];
|
|
l_tmp[3] = product[6] & 0xffffffff00000000ull;
|
|
l_carry -= vli_sub(result, result, l_tmp);
|
|
|
|
if (l_carry < 0)
|
|
{
|
|
do
|
|
{
|
|
l_carry += vli_add(result, result, g_curve_p);
|
|
}
|
|
while (l_carry < 0);
|
|
}
|
|
else
|
|
{
|
|
while (l_carry || vli_cmp(g_curve_p, result) != 1)
|
|
{
|
|
l_carry -= vli_sub(result, result, g_curve_p);
|
|
}
|
|
}
|
|
}
|
|
|
|
#elif ECC_CURVE == secp384r1
|
|
|
|
static void omega_mult(uint64_t *result, uint64_t *right)
|
|
{
|
|
uint64_t l_tmp[NUM_ECC_DIGITS];
|
|
uint64_t l_carry;
|
|
uint64_t l_diff;
|
|
uint i;
|
|
|
|
/* Multiply by (2^128 + 2^96 - 2^32 + 1). */
|
|
|
|
vli_set(result, right); /* 1 */
|
|
l_carry = vli_lshift(l_tmp, right, 32);
|
|
result[1 + NUM_ECC_DIGITS] = l_carry + vli_add(result + 1, result + 1, l_tmp); /* 2^96 + 1 */
|
|
|
|
/* 2^128 + 2^96 + 1 */
|
|
|
|
result[2 + NUM_ECC_DIGITS] = vli_add(result + 2, result + 2, right);
|
|
|
|
/* 2^128 + 2^96 - 2^32 + 1 */
|
|
|
|
l_carry += vli_sub(result, result, l_tmp);
|
|
l_diff = result[NUM_ECC_DIGITS] - l_carry;
|
|
if (l_diff > result[NUM_ECC_DIGITS])
|
|
{
|
|
/* Propagate borrow if necessary. */
|
|
|
|
for (i = 1 + NUM_ECC_DIGITS; ; ++i)
|
|
{
|
|
--result[i];
|
|
if (result[i] != (uint64_t)-1)
|
|
{
|
|
break;
|
|
}
|
|
}
|
|
}
|
|
|
|
result[NUM_ECC_DIGITS] = l_diff;
|
|
}
|
|
|
|
/* Computes result = product % g_curve_p
|
|
* see PDF "Comparing Elliptic Curve Cryptography and RSA on 8-bit CPUs"
|
|
* section "Curve-Specific Optimizations"
|
|
*/
|
|
|
|
static void vli_mmod_fast(uint64_t *result, uint64_t *product)
|
|
{
|
|
uint64_t l_tmp[2 * NUM_ECC_DIGITS];
|
|
uint64_t l_carry;
|
|
uint64_t l_sum;
|
|
uint i;
|
|
|
|
while (!vli_iszero(product + NUM_ECC_DIGITS)) /* While c1 != 0 */
|
|
{
|
|
l_carry = 0;
|
|
|
|
vli_clear(l_tmp);
|
|
vli_clear(l_tmp + NUM_ECC_DIGITS);
|
|
omega_mult(l_tmp, product + NUM_ECC_DIGITS); /* tmp = w * c1 */
|
|
|
|
/* p = c0 */
|
|
|
|
vli_clear(product + NUM_ECC_DIGITS);
|
|
|
|
/* (c1, c0) = c0 + w * c1 */
|
|
|
|
for (i = 0; i < NUM_ECC_DIGITS + 3; ++i)
|
|
{
|
|
l_sum = product[i] + l_tmp[i] + l_carry;
|
|
if (l_sum != product[i])
|
|
{
|
|
l_carry = (l_sum < product[i]);
|
|
}
|
|
|
|
product[i] = l_sum;
|
|
}
|
|
}
|
|
|
|
while (vli_cmp(product, g_curve_p) > 0)
|
|
{
|
|
vli_sub(product, product, g_curve_p);
|
|
}
|
|
|
|
vli_set(result, product);
|
|
}
|
|
|
|
#endif
|
|
|
|
/* Computes result = (left * right) % g_curve_p. */
|
|
|
|
static void vli_modmult_fast(FAR uint64_t *result, FAR uint64_t *left,
|
|
FAR uint64_t *right)
|
|
{
|
|
uint64_t l_product[2 * NUM_ECC_DIGITS];
|
|
|
|
vli_mult(l_product, left, right);
|
|
vli_mmod_fast(result, l_product);
|
|
}
|
|
|
|
/* Computes result = left^2 % g_curve_p. */
|
|
|
|
static void vli_modsquare_fast(FAR uint64_t *result,
|
|
FAR uint64_t *left)
|
|
{
|
|
uint64_t l_product[2 * NUM_ECC_DIGITS];
|
|
|
|
vli_square(l_product, left);
|
|
vli_mmod_fast(result, l_product);
|
|
}
|
|
|
|
/* Computes result = (1 / input) % mod. All VLIs are the same size.
|
|
* See "From Euclid's GCD to Montgomery Multiplication to the Great Divide"
|
|
* https://labs.oracle.com/techrep/2001/smli_tr-2001-95.pdf
|
|
*/
|
|
|
|
static void vli_modinv(FAR uint64_t *result, FAR uint64_t *input,
|
|
FAR uint64_t *mod)
|
|
{
|
|
uint64_t a[NUM_ECC_DIGITS];
|
|
uint64_t b[NUM_ECC_DIGITS];
|
|
uint64_t u[NUM_ECC_DIGITS];
|
|
uint64_t v[NUM_ECC_DIGITS];
|
|
uint64_t l_carry;
|
|
int l_cmpresult;
|
|
|
|
if (vli_iszero(input))
|
|
{
|
|
vli_clear(result);
|
|
return;
|
|
}
|
|
|
|
vli_set(a, input);
|
|
vli_set(b, mod);
|
|
vli_clear(u);
|
|
u[0] = 1;
|
|
vli_clear(v);
|
|
|
|
while ((l_cmpresult = vli_cmp(a, b)) != 0)
|
|
{
|
|
l_carry = 0;
|
|
if (EVEN(a))
|
|
{
|
|
vli_rshift1(a);
|
|
if (!EVEN(u))
|
|
{
|
|
l_carry = vli_add(u, u, mod);
|
|
}
|
|
|
|
vli_rshift1(u);
|
|
if (l_carry)
|
|
{
|
|
u[NUM_ECC_DIGITS - 1] |= 0x8000000000000000ull;
|
|
}
|
|
}
|
|
else if (EVEN(b))
|
|
{
|
|
vli_rshift1(b);
|
|
if (!EVEN(v))
|
|
{
|
|
l_carry = vli_add(v, v, mod);
|
|
}
|
|
|
|
vli_rshift1(v);
|
|
if (l_carry)
|
|
{
|
|
v[NUM_ECC_DIGITS - 1] |= 0x8000000000000000ull;
|
|
}
|
|
}
|
|
else if (l_cmpresult > 0)
|
|
{
|
|
vli_sub(a, a, b);
|
|
vli_rshift1(a);
|
|
if (vli_cmp(u, v) < 0)
|
|
{
|
|
vli_add(u, u, mod);
|
|
}
|
|
|
|
vli_sub(u, u, v);
|
|
if (!EVEN(u))
|
|
{
|
|
l_carry = vli_add(u, u, mod);
|
|
}
|
|
|
|
vli_rshift1(u);
|
|
if (l_carry)
|
|
{
|
|
u[NUM_ECC_DIGITS - 1] |= 0x8000000000000000ull;
|
|
}
|
|
}
|
|
else
|
|
{
|
|
vli_sub(b, b, a);
|
|
vli_rshift1(b);
|
|
if (vli_cmp(v, u) < 0)
|
|
{
|
|
vli_add(v, v, mod);
|
|
}
|
|
|
|
vli_sub(v, v, u);
|
|
if (!EVEN(v))
|
|
{
|
|
l_carry = vli_add(v, v, mod);
|
|
}
|
|
|
|
vli_rshift1(v);
|
|
if (l_carry)
|
|
{
|
|
v[NUM_ECC_DIGITS - 1] |= 0x8000000000000000ull;
|
|
}
|
|
}
|
|
}
|
|
|
|
vli_set(result, u);
|
|
}
|
|
|
|
/* ------ Point operations ------ */
|
|
|
|
/* Returns 1 if point is the point at infinity, 0 otherwise. */
|
|
|
|
static int eccpoint_iszero(FAR eccpoint_t *point)
|
|
{
|
|
return vli_iszero(point->x) && vli_iszero(point->y);
|
|
}
|
|
|
|
/* Point multiplication algorithm using Montgomery's ladder with
|
|
* co-Z coordinates. From http://eprint.iacr.org/2011/338.pdf
|
|
*/
|
|
|
|
/* Double in place */
|
|
|
|
static void eccpoint_double_jacobian(FAR uint64_t *X1,
|
|
FAR uint64_t *Y1,
|
|
FAR uint64_t *Z1)
|
|
{
|
|
/* t1 = X, t2 = Y, t3 = Z */
|
|
|
|
uint64_t t4[NUM_ECC_DIGITS];
|
|
uint64_t t5[NUM_ECC_DIGITS];
|
|
uint64_t l_carry;
|
|
|
|
if (vli_iszero(Z1))
|
|
{
|
|
return;
|
|
}
|
|
|
|
vli_modsquare_fast(t4, Y1); /* t4 = y1^2 */
|
|
vli_modmult_fast(t5, X1, t4); /* t5 = x1*y1^2 = A */
|
|
vli_modsquare_fast(t4, t4); /* t4 = y1^4 */
|
|
vli_modmult_fast(Y1, Y1, Z1); /* t2 = y1*z1 = z3 */
|
|
vli_modsquare_fast(Z1, Z1); /* t3 = z1^2 */
|
|
|
|
vli_modadd(X1, X1, Z1, g_curve_p); /* t1 = x1 + z1^2 */
|
|
vli_modadd(Z1, Z1, Z1, g_curve_p); /* t3 = 2*z1^2 */
|
|
vli_modsub(Z1, X1, Z1, g_curve_p); /* t3 = x1 - z1^2 */
|
|
|
|
/* t1 = x1^2 - z1^4 */
|
|
|
|
vli_modmult_fast(X1, X1, Z1);
|
|
|
|
vli_modadd(Z1, X1, X1, g_curve_p); /* t3 = 2*(x1^2 - z1^4) */
|
|
vli_modadd(X1, X1, Z1, g_curve_p); /* t1 = 3*(x1^2 - z1^4) */
|
|
if (vli_testbit(X1, 0))
|
|
{
|
|
l_carry = vli_add(X1, X1, g_curve_p);
|
|
vli_rshift1(X1);
|
|
X1[NUM_ECC_DIGITS - 1] |= l_carry << 63;
|
|
}
|
|
else
|
|
{
|
|
vli_rshift1(X1);
|
|
}
|
|
|
|
/* t1 = 3/2*(x1^2 - z1^4) = B */
|
|
|
|
/* t3 = B^2 */
|
|
|
|
vli_modsquare_fast(Z1, X1);
|
|
|
|
/* t3 = B^2 - A */
|
|
|
|
vli_modsub(Z1, Z1, t5, g_curve_p);
|
|
|
|
/* t3 = B^2 - 2A = x3 */
|
|
|
|
vli_modsub(Z1, Z1, t5, g_curve_p);
|
|
|
|
/* t5 = A - x3 */
|
|
|
|
vli_modsub(t5, t5, Z1, g_curve_p);
|
|
|
|
/* t1 = B * (A - x3) */
|
|
|
|
vli_modmult_fast(X1, X1, t5);
|
|
|
|
/* t4 = B * (A - x3) - y1^4 = y3 */
|
|
|
|
vli_modsub(t4, X1, t4, g_curve_p);
|
|
|
|
vli_set(X1, Z1);
|
|
vli_set(Z1, Y1);
|
|
vli_set(Y1, t4);
|
|
}
|
|
|
|
/* Modify (x1, y1) => (x1 * z^2, y1 * z^3) */
|
|
|
|
static void apply_z(FAR uint64_t *X1, FAR uint64_t *Y1,
|
|
FAR uint64_t *Z)
|
|
{
|
|
uint64_t t1[NUM_ECC_DIGITS];
|
|
|
|
/* z^2 */
|
|
|
|
vli_modsquare_fast(t1, Z);
|
|
|
|
/* x1 * z^2 */
|
|
|
|
vli_modmult_fast(X1, X1, t1);
|
|
|
|
/* z^3 */
|
|
|
|
vli_modmult_fast(t1, t1, Z);
|
|
|
|
/* y1 * z^3 */
|
|
|
|
vli_modmult_fast(Y1, Y1, t1);
|
|
}
|
|
|
|
/* P = (x1, y1) => 2P, (x2, y2) => P' */
|
|
|
|
static void xycz_initial_double(FAR uint64_t *X1, FAR uint64_t *Y1,
|
|
FAR uint64_t *X2, FAR uint64_t *Y2,
|
|
FAR uint64_t *initialz)
|
|
{
|
|
uint64_t z[NUM_ECC_DIGITS];
|
|
|
|
vli_set(X2, X1);
|
|
vli_set(Y2, Y1);
|
|
|
|
vli_clear(z);
|
|
z[0] = 1;
|
|
if (initialz)
|
|
{
|
|
vli_set(z, initialz);
|
|
}
|
|
|
|
apply_z(X1, Y1, z);
|
|
|
|
eccpoint_double_jacobian(X1, Y1, z);
|
|
|
|
apply_z(X2, Y2, z);
|
|
}
|
|
|
|
/* Input P = (x1, y1, Z), Q = (x2, y2, Z)
|
|
* Output P' = (x1', y1', Z3), P + Q = (x3, y3, Z3)
|
|
* or P => P', Q => P + Q
|
|
*/
|
|
|
|
static void xycz_add(FAR uint64_t *X1, FAR uint64_t *Y1,
|
|
FAR uint64_t *X2, FAR uint64_t *Y2)
|
|
{
|
|
/* t1 = X1, t2 = Y1, t3 = X2, t4 = Y2 */
|
|
|
|
uint64_t t5[NUM_ECC_DIGITS];
|
|
|
|
/* t5 = x2 - x1 */
|
|
|
|
vli_modsub(t5, X2, X1, g_curve_p);
|
|
|
|
/* t5 = (x2 - x1)^2 = A */
|
|
|
|
vli_modsquare_fast(t5, t5);
|
|
|
|
/* t1 = x1*A = B */
|
|
|
|
vli_modmult_fast(X1, X1, t5);
|
|
|
|
/* t3 = x2*A = C */
|
|
|
|
vli_modmult_fast(X2, X2, t5);
|
|
|
|
/* t4 = y2 - y1 */
|
|
|
|
vli_modsub(Y2, Y2, Y1, g_curve_p);
|
|
|
|
/* t5 = (y2 - y1)^2 = D */
|
|
|
|
vli_modsquare_fast(t5, Y2);
|
|
|
|
/* t5 = D - B */
|
|
|
|
vli_modsub(t5, t5, X1, g_curve_p);
|
|
|
|
/* t5 = D - B - C = x3 */
|
|
|
|
vli_modsub(t5, t5, X2, g_curve_p);
|
|
|
|
/* t3 = C - B */
|
|
|
|
vli_modsub(X2, X2, X1, g_curve_p);
|
|
|
|
/* t2 = y1*(C - B) */
|
|
|
|
vli_modmult_fast(Y1, Y1, X2);
|
|
|
|
/* t3 = B - x3 */
|
|
|
|
vli_modsub(X2, X1, t5, g_curve_p);
|
|
|
|
/* t4 = (y2 - y1)*(B - x3) */
|
|
|
|
vli_modmult_fast(Y2, Y2, X2);
|
|
|
|
/* t4 = y3 */
|
|
|
|
vli_modsub(Y2, Y2, Y1, g_curve_p);
|
|
vli_set(X2, t5);
|
|
}
|
|
|
|
/* Input P = (x1, y1, Z), Q = (x2, y2, Z)
|
|
* Output P + Q = (x3, y3, Z3), P - Q = (x3', y3', Z3)
|
|
* or P => P - Q, Q => P + Q
|
|
*/
|
|
|
|
static void xycz_addc(FAR uint64_t *X1, FAR uint64_t *Y1,
|
|
FAR uint64_t *X2, FAR uint64_t *Y2)
|
|
{
|
|
/* t1 = X1, t2 = Y1, t3 = X2, t4 = Y2 */
|
|
|
|
uint64_t t5[NUM_ECC_DIGITS];
|
|
uint64_t t6[NUM_ECC_DIGITS];
|
|
uint64_t t7[NUM_ECC_DIGITS];
|
|
|
|
/* t5 = x2 - x1 */
|
|
|
|
vli_modsub(t5, X2, X1, g_curve_p);
|
|
|
|
/* t5 = (x2 - x1)^2 = A */
|
|
|
|
vli_modsquare_fast(t5, t5);
|
|
|
|
/* t1 = x1*A = B */
|
|
|
|
vli_modmult_fast(X1, X1, t5);
|
|
|
|
/* t3 = x2*A = C */
|
|
|
|
vli_modmult_fast(X2, X2, t5);
|
|
|
|
/* t4 = y2 + y1 */
|
|
|
|
vli_modadd(t5, Y2, Y1, g_curve_p);
|
|
|
|
/* t4 = y2 - y1 */
|
|
|
|
vli_modsub(Y2, Y2, Y1, g_curve_p);
|
|
|
|
/* t6 = C - B */
|
|
|
|
vli_modsub(t6, X2, X1, g_curve_p);
|
|
|
|
/* t2 = y1 * (C - B) */
|
|
|
|
vli_modmult_fast(Y1, Y1, t6);
|
|
|
|
/* t6 = B + C */
|
|
|
|
vli_modadd(t6, X1, X2, g_curve_p);
|
|
|
|
/* t3 = (y2 - y1)^2 */
|
|
|
|
vli_modsquare_fast(X2, Y2);
|
|
|
|
/* t3 = x3 */
|
|
|
|
vli_modsub(X2, X2, t6, g_curve_p);
|
|
|
|
/* t7 = B - x3 */
|
|
|
|
vli_modsub(t7, X1, X2, g_curve_p);
|
|
|
|
/* t4 = (y2 - y1)*(B - x3) */
|
|
|
|
vli_modmult_fast(Y2, Y2, t7);
|
|
|
|
/* t4 = y3 */
|
|
|
|
vli_modsub(Y2, Y2, Y1, g_curve_p);
|
|
|
|
/* t7 = (y2 + y1)^2 = F */
|
|
|
|
vli_modsquare_fast(t7, t5);
|
|
|
|
/* t7 = x3' */
|
|
|
|
vli_modsub(t7, t7, t6, g_curve_p);
|
|
|
|
/* t6 = x3' - B */
|
|
|
|
vli_modsub(t6, t7, X1, g_curve_p);
|
|
|
|
/* t6 = (y2 + y1)*(x3' - B) */
|
|
|
|
vli_modmult_fast(t6, t6, t5);
|
|
|
|
/* t2 = y3' */
|
|
|
|
vli_modsub(Y1, t6, Y1, g_curve_p);
|
|
|
|
vli_set(X1, t7);
|
|
}
|
|
|
|
static void eccpoint_mult(FAR eccpoint_t *result, FAR eccpoint_t *point,
|
|
FAR uint64_t *scalar, FAR uint64_t *initialz)
|
|
{
|
|
/* R0 and R1 */
|
|
|
|
uint64_t rx[2][NUM_ECC_DIGITS];
|
|
uint64_t ry[2][NUM_ECC_DIGITS];
|
|
uint64_t z[NUM_ECC_DIGITS];
|
|
int nb;
|
|
int i;
|
|
|
|
vli_set(rx[1], point->x);
|
|
vli_set(ry[1], point->y);
|
|
|
|
xycz_initial_double(rx[1], ry[1], rx[0], ry[0], initialz);
|
|
|
|
for (i = vli_numbits(scalar) - 2; i > 0; --i)
|
|
{
|
|
nb = !vli_testbit(scalar, i);
|
|
xycz_addc(rx[1 - nb], ry[1 - nb], rx[nb], ry[nb]);
|
|
xycz_add(rx[nb], ry[nb], rx[1 - nb], ry[1 - nb]);
|
|
}
|
|
|
|
nb = !vli_testbit(scalar, 0);
|
|
xycz_addc(rx[1 - nb], ry[1 - nb], rx[nb], ry[nb]);
|
|
|
|
/* Find final 1/Z value. */
|
|
|
|
/* X1 - X0 */
|
|
|
|
vli_modsub(z, rx[1], rx[0], g_curve_p);
|
|
|
|
/* Yb * (X1 - X0) */
|
|
|
|
vli_modmult_fast(z, z, ry[1 - nb]);
|
|
|
|
/* xP * Yb * (X1 - X0) */
|
|
|
|
vli_modmult_fast(z, z, point->x);
|
|
|
|
/* 1 / (xP * Yb * (X1 - X0)) */
|
|
|
|
vli_modinv(z, z, g_curve_p);
|
|
|
|
/* yP / (xP * Yb * (X1 - X0)) */
|
|
|
|
vli_modmult_fast(z, z, point->y);
|
|
|
|
/* Xb * yP / (xP * Yb * (X1 - X0)) */
|
|
|
|
vli_modmult_fast(z, z, rx[1 - nb]);
|
|
|
|
/* End 1/Z calculation */
|
|
|
|
xycz_add(rx[nb], ry[nb], rx[1 - nb], ry[1 - nb]);
|
|
|
|
apply_z(rx[0], ry[0], z);
|
|
|
|
vli_set(result->x, rx[0]);
|
|
vli_set(result->y, ry[0]);
|
|
}
|
|
|
|
static void ecc_bytes2native(uint64_t native[NUM_ECC_DIGITS],
|
|
const uint8_t bytes[ECC_BYTES])
|
|
{
|
|
FAR const uint8_t *digit;
|
|
unsigned i;
|
|
|
|
for (i = 0; i < NUM_ECC_DIGITS; ++i)
|
|
{
|
|
digit = bytes + 8 * (NUM_ECC_DIGITS - 1 - i);
|
|
native[i] = ((uint64_t)digit[0] << 56) | ((uint64_t)digit[1] << 48) |
|
|
((uint64_t)digit[2] << 40) | ((uint64_t)digit[3] << 32) |
|
|
((uint64_t)digit[4] << 24) | ((uint64_t)digit[5] << 16) |
|
|
((uint64_t)digit[6] << 8) | (uint64_t)digit[7];
|
|
}
|
|
}
|
|
|
|
static void ecc_native2bytes(uint8_t bytes[ECC_BYTES],
|
|
const uint64_t native[NUM_ECC_DIGITS])
|
|
{
|
|
FAR uint8_t *digit;
|
|
unsigned i;
|
|
|
|
for (i = 0; i < NUM_ECC_DIGITS; ++i)
|
|
{
|
|
digit = bytes + 8 * (NUM_ECC_DIGITS - 1 - i);
|
|
digit[0] = native[i] >> 56;
|
|
digit[1] = native[i] >> 48;
|
|
digit[2] = native[i] >> 40;
|
|
digit[3] = native[i] >> 32;
|
|
digit[4] = native[i] >> 24;
|
|
digit[5] = native[i] >> 16;
|
|
digit[6] = native[i] >> 8;
|
|
digit[7] = native[i];
|
|
}
|
|
}
|
|
|
|
/* Compute a = sqrt(a) (mod g_curve_p). */
|
|
|
|
static void mod_sqrt(uint64_t a[NUM_ECC_DIGITS])
|
|
{
|
|
unsigned i;
|
|
uint64_t l_result[NUM_ECC_DIGITS] =
|
|
{
|
|
1
|
|
};
|
|
|
|
uint64_t p1[NUM_ECC_DIGITS] =
|
|
{
|
|
1
|
|
};
|
|
|
|
/* Since g_curve_p == 3 (mod 4) for all supported curves, we can
|
|
* compute sqrt(a) = a^((g_curve_p + 1) / 4) (mod g_curve_p).
|
|
*/
|
|
|
|
vli_add(p1, g_curve_p, p1); /* p1 = g_curve_p + 1 */
|
|
for (i = vli_numbits(p1) - 1; i > 1; --i)
|
|
{
|
|
vli_modsquare_fast(l_result, l_result);
|
|
if (vli_testbit(p1, i))
|
|
{
|
|
vli_modmult_fast(l_result, l_result, a);
|
|
}
|
|
}
|
|
|
|
vli_set(a, l_result);
|
|
}
|
|
|
|
static void
|
|
ecc_point_decompress(FAR eccpoint_t *point,
|
|
const uint8_t compressed[ECC_BYTES + 1])
|
|
{
|
|
/* -a = 3 */
|
|
|
|
uint64_t _3[NUM_ECC_DIGITS] =
|
|
{
|
|
3
|
|
};
|
|
|
|
ecc_bytes2native(point->x, compressed + 1);
|
|
|
|
/* y = x^2 */
|
|
|
|
vli_modsquare_fast(point->y, point->x);
|
|
|
|
/* y = x^2 - 3 */
|
|
|
|
vli_modsub(point->y, point->y, _3, g_curve_p);
|
|
|
|
/* y = x^3 - 3x */
|
|
|
|
vli_modmult_fast(point->y, point->y, point->x);
|
|
|
|
/* y = x^3 - 3x + b */
|
|
|
|
vli_modadd(point->y, point->y, g_curve_b, g_curve_p);
|
|
|
|
mod_sqrt(point->y);
|
|
|
|
if ((point->y[0] & 0x01) != (compressed[0] & 0x01))
|
|
{
|
|
vli_sub(point->y, g_curve_p, point->y);
|
|
}
|
|
}
|
|
|
|
/* -------- ECDSA code -------- */
|
|
|
|
/* Computes result = (left * right) % mod. */
|
|
|
|
static void vli_modmult(FAR uint64_t *result, FAR uint64_t *left,
|
|
FAR uint64_t *right, FAR uint64_t *mod)
|
|
{
|
|
uint64_t l_product[2 * NUM_ECC_DIGITS];
|
|
uint64_t l_modmultiple[2 * NUM_ECC_DIGITS];
|
|
uint64_t l_carry;
|
|
uint l_modbits = vli_numbits(mod);
|
|
uint l_productbits;
|
|
uint l_digitshift;
|
|
uint l_bitshift;
|
|
int l_cmp;
|
|
|
|
vli_mult(l_product, left, right);
|
|
l_productbits = vli_numbits(l_product + NUM_ECC_DIGITS);
|
|
if (l_productbits)
|
|
{
|
|
l_productbits += NUM_ECC_DIGITS * 64;
|
|
}
|
|
else
|
|
{
|
|
l_productbits = vli_numbits(l_product);
|
|
}
|
|
|
|
if (l_productbits < l_modbits)
|
|
{
|
|
/* l_product < mod. */
|
|
|
|
vli_set(result, l_product);
|
|
return;
|
|
}
|
|
|
|
/* Shift mod by (l_leftBits - l_modbits).
|
|
* This multiplies mod by the largest power of two possible
|
|
* while still resulting in a number less than left.
|
|
*/
|
|
|
|
vli_clear(l_modmultiple);
|
|
vli_clear(l_modmultiple + NUM_ECC_DIGITS);
|
|
l_digitshift = (l_productbits - l_modbits) / 64;
|
|
l_bitshift = (l_productbits - l_modbits) % 64;
|
|
if (l_bitshift)
|
|
{
|
|
l_modmultiple[l_digitshift + NUM_ECC_DIGITS] =
|
|
vli_lshift(l_modmultiple + l_digitshift, mod, l_bitshift);
|
|
}
|
|
else
|
|
{
|
|
vli_set(l_modmultiple + l_digitshift, mod);
|
|
}
|
|
|
|
/* Subtract all multiples of mod to get the remainder. */
|
|
|
|
vli_clear(result);
|
|
|
|
/* Use result as a temp var to store 1 (for subtraction) */
|
|
|
|
result[0] = 1;
|
|
while (l_productbits > NUM_ECC_DIGITS * 64 ||
|
|
vli_cmp(l_modmultiple, mod) >= 0)
|
|
{
|
|
l_cmp = vli_cmp(l_modmultiple + NUM_ECC_DIGITS,
|
|
l_product + NUM_ECC_DIGITS);
|
|
if (l_cmp < 0 ||
|
|
(l_cmp == 0 && vli_cmp(l_modmultiple, l_product) <= 0))
|
|
{
|
|
if (vli_sub(l_product, l_product, l_modmultiple))
|
|
{
|
|
/* borrow */
|
|
|
|
vli_sub(l_product + NUM_ECC_DIGITS,
|
|
l_product + NUM_ECC_DIGITS, result);
|
|
}
|
|
|
|
vli_sub(l_product + NUM_ECC_DIGITS, l_product + NUM_ECC_DIGITS,
|
|
l_modmultiple + NUM_ECC_DIGITS);
|
|
}
|
|
|
|
l_carry = (l_modmultiple[NUM_ECC_DIGITS] & 0x01) << 63;
|
|
vli_rshift1(l_modmultiple + NUM_ECC_DIGITS);
|
|
vli_rshift1(l_modmultiple);
|
|
l_modmultiple[NUM_ECC_DIGITS - 1] |= l_carry;
|
|
--l_productbits;
|
|
}
|
|
|
|
vli_set(result, l_product);
|
|
}
|
|
|
|
static uint umax(uint a, uint b)
|
|
{
|
|
return a > b ? a : b;
|
|
}
|
|
|
|
/****************************************************************************
|
|
* Public Functions
|
|
****************************************************************************/
|
|
|
|
int ecc_make_key(uint8_t publickey[ECC_BYTES + 1],
|
|
uint8_t privatekey[ECC_BYTES])
|
|
{
|
|
uint64_t l_private[NUM_ECC_DIGITS];
|
|
eccpoint_t l_public;
|
|
unsigned l_tries = 0;
|
|
|
|
memset(&l_public, 0, sizeof(eccpoint_t));
|
|
|
|
do
|
|
{
|
|
if (l_tries++ >= MAX_TRIES)
|
|
{
|
|
return 0;
|
|
}
|
|
|
|
arc4random_buf(l_private, NUM_ECC_DIGITS);
|
|
|
|
if (vli_iszero(l_private))
|
|
{
|
|
continue;
|
|
}
|
|
|
|
/* Make sure the private key is in the range [1, n-1].
|
|
* For the supported curves, n is always large enough that we only
|
|
* need to subtract once at most.
|
|
*/
|
|
|
|
if (vli_cmp(g_curve_n, l_private) != 1)
|
|
{
|
|
vli_sub(l_private, l_private, g_curve_n);
|
|
}
|
|
|
|
eccpoint_mult(&l_public, &g_curve_g, l_private, NULL);
|
|
}
|
|
while (eccpoint_iszero(&l_public));
|
|
|
|
ecc_native2bytes(privatekey, l_private);
|
|
ecc_native2bytes(publickey + 1, l_public.x);
|
|
publickey[0] = 2 + (l_public.y[0] & 0x01);
|
|
return 1;
|
|
}
|
|
|
|
int ecc_make_key_uncomp(uint8_t publickey_x[ECC_BYTES],
|
|
uint8_t publickey_y[ECC_BYTES],
|
|
uint8_t privatekey[ECC_BYTES])
|
|
{
|
|
uint64_t l_private[NUM_ECC_DIGITS];
|
|
eccpoint_t l_public;
|
|
unsigned l_tries = 0;
|
|
|
|
do
|
|
{
|
|
if (l_tries++ >= MAX_TRIES)
|
|
{
|
|
return 0;
|
|
}
|
|
|
|
arc4random_buf(l_private, NUM_ECC_DIGITS);
|
|
|
|
if (vli_iszero(l_private))
|
|
{
|
|
continue;
|
|
}
|
|
|
|
/* Make sure the private key is in the range [1, n-1].
|
|
* For the supported curves, n is always large enough that we only
|
|
* need to subtract once at most.
|
|
*/
|
|
|
|
if (vli_cmp(g_curve_n, l_private) != 1)
|
|
{
|
|
vli_sub(l_private, l_private, g_curve_n);
|
|
}
|
|
|
|
eccpoint_mult(&l_public, &g_curve_g, l_private, NULL);
|
|
}
|
|
while (eccpoint_iszero(&l_public));
|
|
|
|
ecc_native2bytes(privatekey, l_private);
|
|
ecc_native2bytes(publickey_x, l_public.x);
|
|
ecc_native2bytes(publickey_y, l_public.y);
|
|
return 1;
|
|
}
|
|
|
|
int ecdh_shared_secret(const uint8_t publickey[ECC_BYTES + 1],
|
|
const uint8_t privatekey[ECC_BYTES],
|
|
uint8_t secret[ECC_BYTES])
|
|
{
|
|
eccpoint_t l_product;
|
|
eccpoint_t l_public;
|
|
uint64_t l_private[NUM_ECC_DIGITS];
|
|
uint64_t l_random[NUM_ECC_DIGITS];
|
|
|
|
memset(&l_product, 0, sizeof(eccpoint_t));
|
|
memset(&l_public, 0, sizeof(eccpoint_t));
|
|
|
|
arc4random_buf(l_random, NUM_ECC_DIGITS);
|
|
ecc_point_decompress(&l_public, publickey);
|
|
ecc_bytes2native(l_private, privatekey);
|
|
|
|
eccpoint_mult(&l_product, &l_public, l_private, l_random);
|
|
|
|
ecc_native2bytes(secret, l_product.x);
|
|
|
|
return !eccpoint_iszero(&l_product);
|
|
}
|
|
|
|
int ecdsa_sign(const uint8_t privatekey[ECC_BYTES],
|
|
const uint8_t hash[ECC_BYTES],
|
|
uint8_t signature[ECC_BYTES * 2])
|
|
{
|
|
uint64_t k[NUM_ECC_DIGITS];
|
|
uint64_t l_tmp[NUM_ECC_DIGITS];
|
|
uint64_t l_s[NUM_ECC_DIGITS];
|
|
unsigned l_tries = 0;
|
|
eccpoint_t p;
|
|
|
|
memset(&p, 0, sizeof(eccpoint_t));
|
|
|
|
do
|
|
{
|
|
if (l_tries++ >= MAX_TRIES)
|
|
{
|
|
return 0;
|
|
}
|
|
|
|
arc4random_buf(k, NUM_ECC_DIGITS);
|
|
|
|
if (vli_iszero(k))
|
|
{
|
|
continue;
|
|
}
|
|
|
|
if (vli_cmp(g_curve_n, k) != 1)
|
|
{
|
|
vli_sub(k, k, g_curve_n);
|
|
}
|
|
|
|
/* tmp = k * G */
|
|
|
|
eccpoint_mult(&p, &g_curve_g, k, NULL);
|
|
|
|
/* r = x1 (mod n) */
|
|
|
|
if (vli_cmp(g_curve_n, p.x) != 1)
|
|
{
|
|
vli_sub(p.x, p.x, g_curve_n);
|
|
}
|
|
}
|
|
while (vli_iszero(p.x));
|
|
|
|
ecc_native2bytes(signature, p.x);
|
|
|
|
ecc_bytes2native(l_tmp, privatekey);
|
|
vli_modmult(l_s, p.x, l_tmp, g_curve_n); /* s = r*d */
|
|
ecc_bytes2native(l_tmp, hash);
|
|
vli_modadd(l_s, l_tmp, l_s, g_curve_n); /* s = e + r*d */
|
|
|
|
/* k = 1 / k */
|
|
|
|
vli_modinv(k, k, g_curve_n);
|
|
|
|
/* s = (e + r*d) / k */
|
|
|
|
vli_modmult(l_s, l_s, k, g_curve_n);
|
|
ecc_native2bytes(signature + ECC_BYTES, l_s);
|
|
|
|
return 1;
|
|
}
|
|
|
|
int ecdsa_verify(const uint8_t publickey[ECC_BYTES + 1],
|
|
const uint8_t hash[ECC_BYTES],
|
|
const uint8_t signature[ECC_BYTES * 2])
|
|
{
|
|
uint64_t u1[NUM_ECC_DIGITS];
|
|
uint64_t u2[NUM_ECC_DIGITS];
|
|
uint64_t z[NUM_ECC_DIGITS];
|
|
uint64_t rx[NUM_ECC_DIGITS];
|
|
uint64_t ry[NUM_ECC_DIGITS];
|
|
uint64_t tx[NUM_ECC_DIGITS];
|
|
uint64_t ty[NUM_ECC_DIGITS];
|
|
uint64_t tz[NUM_ECC_DIGITS];
|
|
uint64_t l_r[NUM_ECC_DIGITS];
|
|
uint64_t l_s[NUM_ECC_DIGITS];
|
|
uint l_numbits;
|
|
eccpoint_t *l_point;
|
|
eccpoint_t l_public;
|
|
eccpoint_t l_sum;
|
|
int l_index;
|
|
int i;
|
|
|
|
/* Use Shamir's trick to calculate u1*G + u2*Q */
|
|
|
|
eccpoint_t *l_points[4] =
|
|
{
|
|
NULL, &g_curve_g, &l_public, &l_sum
|
|
};
|
|
|
|
ecc_point_decompress(&l_public, publickey);
|
|
ecc_bytes2native(l_r, signature);
|
|
ecc_bytes2native(l_s, signature + ECC_BYTES);
|
|
|
|
if (vli_iszero(l_r) || vli_iszero(l_s))
|
|
{
|
|
/* r, s must not be 0. */
|
|
|
|
return 0;
|
|
}
|
|
|
|
if (vli_cmp(g_curve_n, l_r) != 1 || vli_cmp(g_curve_n, l_s) != 1)
|
|
{
|
|
/* r, s must be < n. */
|
|
|
|
return 0;
|
|
}
|
|
|
|
/* Calculate u1 and u2. */
|
|
|
|
vli_modinv(z, l_s, g_curve_n); /* Z = s^-1 */
|
|
ecc_bytes2native(u1, hash);
|
|
vli_modmult(u1, u1, z, g_curve_n); /* u1 = e/s */
|
|
|
|
/* u2 = r/s */
|
|
|
|
vli_modmult(u2, l_r, z, g_curve_n);
|
|
|
|
/* Calculate l_sum = G + Q. */
|
|
|
|
vli_set(l_sum.x, l_public.x);
|
|
vli_set(l_sum.y, l_public.y);
|
|
vli_set(tx, g_curve_g.x);
|
|
vli_set(ty, g_curve_g.y);
|
|
vli_modsub(z, l_sum.x, tx, g_curve_p); /* Z = x2 - x1 */
|
|
xycz_add(tx, ty, l_sum.x, l_sum.y);
|
|
vli_modinv(z, z, g_curve_p); /* Z = 1/Z */
|
|
apply_z(l_sum.x, l_sum.y, z);
|
|
|
|
l_numbits = umax(vli_numbits(u1), vli_numbits(u2));
|
|
|
|
l_point = l_points[(!!vli_testbit(u1, l_numbits - 1)) |
|
|
((!!vli_testbit(u2, l_numbits - 1)) << 1)];
|
|
vli_set(rx, l_point->x);
|
|
vli_set(ry, l_point->y);
|
|
vli_clear(z);
|
|
z[0] = 1;
|
|
|
|
for (i = l_numbits - 2; i >= 0; --i)
|
|
{
|
|
eccpoint_double_jacobian(rx, ry, z);
|
|
|
|
l_index = (!!vli_testbit(u1, i)) | ((!!vli_testbit(u2, i)) << 1);
|
|
l_point = l_points[l_index];
|
|
if (l_point)
|
|
{
|
|
vli_set(tx, l_point->x);
|
|
vli_set(ty, l_point->y);
|
|
apply_z(tx, ty, z);
|
|
vli_modsub(tz, rx, tx, g_curve_p); /* Z = x2 - x1 */
|
|
xycz_add(tx, ty, rx, ry);
|
|
vli_modmult_fast(z, z, tz);
|
|
}
|
|
}
|
|
|
|
vli_modinv(z, z, g_curve_p); /* Z = 1/Z */
|
|
apply_z(rx, ry, z);
|
|
|
|
/* v = x1 (mod n) */
|
|
|
|
if (vli_cmp(g_curve_n, rx) != 1)
|
|
{
|
|
vli_sub(rx, rx, g_curve_n);
|
|
}
|
|
|
|
/* Accept only if v == r. */
|
|
|
|
return vli_cmp(rx, l_r) == 0;
|
|
}
|