rust-secp256k1-unsafe-fast/secp256k1-sys/depend/secp256k1/src/ecdsa_impl.h

316 lines
11 KiB
C
Raw Normal View History

/***********************************************************************
* Copyright (c) 2013-2015 Pieter Wuille *
* Distributed under the MIT software license, see the accompanying *
* file COPYING or https://www.opensource.org/licenses/mit-license.php.*
***********************************************************************/
2015-10-26 14:54:21 +00:00
#ifndef SECP256K1_ECDSA_IMPL_H
#define SECP256K1_ECDSA_IMPL_H
2015-10-26 14:54:21 +00:00
#include "scalar.h"
#include "field.h"
#include "group.h"
#include "ecmult.h"
#include "ecmult_gen.h"
#include "ecdsa.h"
/** Group order for secp256k1 defined as 'n' in "Standards for Efficient Cryptography" (SEC2) 2.7.1
* sage: for t in xrange(1023, -1, -1):
* .. p = 2**256 - 2**32 - t
* .. if p.is_prime():
* .. print '%x'%p
* .. break
* 'fffffffffffffffffffffffffffffffffffffffffffffffffffffffefffffc2f'
* sage: a = 0
* sage: b = 7
* sage: F = FiniteField (p)
* sage: '%x' % (EllipticCurve ([F (a), F (b)]).order())
* 'fffffffffffffffffffffffffffffffebaaedce6af48a03bbfd25e8cd0364141'
*/
static const rustsecp256k1_v0_7_0_fe rustsecp256k1_v0_7_0_ecdsa_const_order_as_fe = SECP256K1_FE_CONST(
2015-10-26 14:54:21 +00:00
0xFFFFFFFFUL, 0xFFFFFFFFUL, 0xFFFFFFFFUL, 0xFFFFFFFEUL,
0xBAAEDCE6UL, 0xAF48A03BUL, 0xBFD25E8CUL, 0xD0364141UL
);
/** Difference between field and order, values 'p' and 'n' values defined in
* "Standards for Efficient Cryptography" (SEC2) 2.7.1.
* sage: p = 0xFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFEFFFFFC2F
* sage: a = 0
* sage: b = 7
* sage: F = FiniteField (p)
* sage: '%x' % (p - EllipticCurve ([F (a), F (b)]).order())
* '14551231950b75fc4402da1722fc9baee'
*/
static const rustsecp256k1_v0_7_0_fe rustsecp256k1_v0_7_0_ecdsa_const_p_minus_order = SECP256K1_FE_CONST(
2015-10-26 14:54:21 +00:00
0, 0, 0, 1, 0x45512319UL, 0x50B75FC4UL, 0x402DA172UL, 0x2FC9BAEEUL
);
static int rustsecp256k1_v0_7_0_der_read_len(size_t *len, const unsigned char **sigp, const unsigned char *sigend) {
size_t lenleft;
unsigned char b1;
VERIFY_CHECK(len != NULL);
*len = 0;
2015-10-26 14:54:21 +00:00
if (*sigp >= sigend) {
return 0;
2015-10-26 14:54:21 +00:00
}
b1 = *((*sigp)++);
if (b1 == 0xFF) {
/* X.690-0207 8.1.3.5.c the value 0xFF shall not be used. */
return 0;
2015-10-26 14:54:21 +00:00
}
if ((b1 & 0x80) == 0) {
/* X.690-0207 8.1.3.4 short form length octets */
*len = b1;
return 1;
2015-10-26 14:54:21 +00:00
}
if (b1 == 0x80) {
/* Indefinite length is not allowed in DER. */
return 0;
2015-10-26 14:54:21 +00:00
}
/* X.690-207 8.1.3.5 long form length octets */
lenleft = b1 & 0x7F; /* lenleft is at least 1 */
if (lenleft > (size_t)(sigend - *sigp)) {
return 0;
2015-10-26 14:54:21 +00:00
}
if (**sigp == 0) {
/* Not the shortest possible length encoding. */
return 0;
2015-10-26 14:54:21 +00:00
}
if (lenleft > sizeof(size_t)) {
/* The resulting length would exceed the range of a size_t, so
* certainly longer than the passed array size.
*/
return 0;
2015-10-26 14:54:21 +00:00
}
while (lenleft > 0) {
*len = (*len << 8) | **sigp;
2015-10-26 14:54:21 +00:00
(*sigp)++;
lenleft--;
}
if (*len > (size_t)(sigend - *sigp)) {
/* Result exceeds the length of the passed array. */
return 0;
}
if (*len < 128) {
2015-10-26 14:54:21 +00:00
/* Not the shortest possible length encoding. */
return 0;
2015-10-26 14:54:21 +00:00
}
return 1;
2015-10-26 14:54:21 +00:00
}
static int rustsecp256k1_v0_7_0_der_parse_integer(rustsecp256k1_v0_7_0_scalar *r, const unsigned char **sig, const unsigned char *sigend) {
2015-10-26 14:54:21 +00:00
int overflow = 0;
unsigned char ra[32] = {0};
size_t rlen;
2015-10-26 14:54:21 +00:00
if (*sig == sigend || **sig != 0x02) {
/* Not a primitive integer (X.690-0207 8.3.1). */
return 0;
}
(*sig)++;
if (rustsecp256k1_v0_7_0_der_read_len(&rlen, sig, sigend) == 0) {
return 0;
}
if (rlen == 0 || rlen > (size_t)(sigend - *sig)) {
2015-10-26 14:54:21 +00:00
/* Exceeds bounds or not at least length 1 (X.690-0207 8.3.1). */
return 0;
}
if (**sig == 0x00 && rlen > 1 && (((*sig)[1]) & 0x80) == 0x00) {
/* Excessive 0x00 padding. */
return 0;
}
if (**sig == 0xFF && rlen > 1 && (((*sig)[1]) & 0x80) == 0x80) {
/* Excessive 0xFF padding. */
return 0;
}
if ((**sig & 0x80) == 0x80) {
/* Negative. */
overflow = 1;
}
/* There is at most one leading zero byte:
* if there were two leading zero bytes, we would have failed and returned 0
* because of excessive 0x00 padding already. */
if (rlen > 0 && **sig == 0) {
/* Skip leading zero byte */
2015-10-26 14:54:21 +00:00
rlen--;
(*sig)++;
}
if (rlen > 32) {
overflow = 1;
}
if (!overflow) {
if (rlen) memcpy(ra + 32 - rlen, *sig, rlen);
rustsecp256k1_v0_7_0_scalar_set_b32(r, ra, &overflow);
2015-10-26 14:54:21 +00:00
}
if (overflow) {
rustsecp256k1_v0_7_0_scalar_set_int(r, 0);
2015-10-26 14:54:21 +00:00
}
(*sig) += rlen;
return 1;
}
static int rustsecp256k1_v0_7_0_ecdsa_sig_parse(rustsecp256k1_v0_7_0_scalar *rr, rustsecp256k1_v0_7_0_scalar *rs, const unsigned char *sig, size_t size) {
2015-10-26 14:54:21 +00:00
const unsigned char *sigend = sig + size;
size_t rlen;
2015-10-26 14:54:21 +00:00
if (sig == sigend || *(sig++) != 0x30) {
/* The encoding doesn't start with a constructed sequence (X.690-0207 8.9.1). */
return 0;
}
if (rustsecp256k1_v0_7_0_der_read_len(&rlen, &sig, sigend) == 0) {
2015-10-26 14:54:21 +00:00
return 0;
}
if (rlen != (size_t)(sigend - sig)) {
/* Tuple exceeds bounds or garage after tuple. */
2015-10-26 14:54:21 +00:00
return 0;
}
if (!rustsecp256k1_v0_7_0_der_parse_integer(rr, &sig, sigend)) {
2015-10-26 14:54:21 +00:00
return 0;
}
if (!rustsecp256k1_v0_7_0_der_parse_integer(rs, &sig, sigend)) {
2015-10-26 14:54:21 +00:00
return 0;
}
if (sig != sigend) {
/* Trailing garbage inside tuple. */
return 0;
}
return 1;
}
static int rustsecp256k1_v0_7_0_ecdsa_sig_serialize(unsigned char *sig, size_t *size, const rustsecp256k1_v0_7_0_scalar* ar, const rustsecp256k1_v0_7_0_scalar* as) {
2015-10-26 14:54:21 +00:00
unsigned char r[33] = {0}, s[33] = {0};
unsigned char *rp = r, *sp = s;
size_t lenR = 33, lenS = 33;
rustsecp256k1_v0_7_0_scalar_get_b32(&r[1], ar);
rustsecp256k1_v0_7_0_scalar_get_b32(&s[1], as);
2015-10-26 14:54:21 +00:00
while (lenR > 1 && rp[0] == 0 && rp[1] < 0x80) { lenR--; rp++; }
while (lenS > 1 && sp[0] == 0 && sp[1] < 0x80) { lenS--; sp++; }
if (*size < 6+lenS+lenR) {
*size = 6 + lenS + lenR;
return 0;
}
*size = 6 + lenS + lenR;
sig[0] = 0x30;
sig[1] = 4 + lenS + lenR;
sig[2] = 0x02;
sig[3] = lenR;
memcpy(sig+4, rp, lenR);
sig[4+lenR] = 0x02;
sig[5+lenR] = lenS;
memcpy(sig+lenR+6, sp, lenS);
return 1;
}
static int rustsecp256k1_v0_7_0_ecdsa_sig_verify(const rustsecp256k1_v0_7_0_scalar *sigr, const rustsecp256k1_v0_7_0_scalar *sigs, const rustsecp256k1_v0_7_0_ge *pubkey, const rustsecp256k1_v0_7_0_scalar *message) {
2015-10-26 14:54:21 +00:00
unsigned char c[32];
rustsecp256k1_v0_7_0_scalar sn, u1, u2;
#if !defined(EXHAUSTIVE_TEST_ORDER)
rustsecp256k1_v0_7_0_fe xr;
#endif
rustsecp256k1_v0_7_0_gej pubkeyj;
rustsecp256k1_v0_7_0_gej pr;
2015-10-26 14:54:21 +00:00
if (rustsecp256k1_v0_7_0_scalar_is_zero(sigr) || rustsecp256k1_v0_7_0_scalar_is_zero(sigs)) {
2015-10-26 14:54:21 +00:00
return 0;
}
rustsecp256k1_v0_7_0_scalar_inverse_var(&sn, sigs);
rustsecp256k1_v0_7_0_scalar_mul(&u1, &sn, message);
rustsecp256k1_v0_7_0_scalar_mul(&u2, &sn, sigr);
rustsecp256k1_v0_7_0_gej_set_ge(&pubkeyj, pubkey);
rustsecp256k1_v0_7_0_ecmult(&pr, &pubkeyj, &u2, &u1);
if (rustsecp256k1_v0_7_0_gej_is_infinity(&pr)) {
2015-10-26 14:54:21 +00:00
return 0;
}
#if defined(EXHAUSTIVE_TEST_ORDER)
{
rustsecp256k1_v0_7_0_scalar computed_r;
rustsecp256k1_v0_7_0_ge pr_ge;
rustsecp256k1_v0_7_0_ge_set_gej(&pr_ge, &pr);
rustsecp256k1_v0_7_0_fe_normalize(&pr_ge.x);
rustsecp256k1_v0_7_0_fe_get_b32(c, &pr_ge.x);
rustsecp256k1_v0_7_0_scalar_set_b32(&computed_r, c, NULL);
return rustsecp256k1_v0_7_0_scalar_eq(sigr, &computed_r);
}
#else
rustsecp256k1_v0_7_0_scalar_get_b32(c, sigr);
rustsecp256k1_v0_7_0_fe_set_b32(&xr, c);
2015-10-26 14:54:21 +00:00
/** We now have the recomputed R point in pr, and its claimed x coordinate (modulo n)
* in xr. Naively, we would extract the x coordinate from pr (requiring a inversion modulo p),
* compute the remainder modulo n, and compare it to xr. However:
*
* xr == X(pr) mod n
* <=> exists h. (xr + h * n < p && xr + h * n == X(pr))
* [Since 2 * n > p, h can only be 0 or 1]
* <=> (xr == X(pr)) || (xr + n < p && xr + n == X(pr))
* [In Jacobian coordinates, X(pr) is pr.x / pr.z^2 mod p]
* <=> (xr == pr.x / pr.z^2 mod p) || (xr + n < p && xr + n == pr.x / pr.z^2 mod p)
* [Multiplying both sides of the equations by pr.z^2 mod p]
* <=> (xr * pr.z^2 mod p == pr.x) || (xr + n < p && (xr + n) * pr.z^2 mod p == pr.x)
*
* Thus, we can avoid the inversion, but we have to check both cases separately.
* rustsecp256k1_v0_7_0_gej_eq_x implements the (xr * pr.z^2 mod p == pr.x) test.
2015-10-26 14:54:21 +00:00
*/
if (rustsecp256k1_v0_7_0_gej_eq_x_var(&xr, &pr)) {
2015-10-26 14:54:21 +00:00
/* xr * pr.z^2 mod p == pr.x, so the signature is valid. */
return 1;
}
if (rustsecp256k1_v0_7_0_fe_cmp_var(&xr, &rustsecp256k1_v0_7_0_ecdsa_const_p_minus_order) >= 0) {
2015-10-26 14:54:21 +00:00
/* xr + n >= p, so we can skip testing the second case. */
return 0;
}
rustsecp256k1_v0_7_0_fe_add(&xr, &rustsecp256k1_v0_7_0_ecdsa_const_order_as_fe);
if (rustsecp256k1_v0_7_0_gej_eq_x_var(&xr, &pr)) {
2015-10-26 14:54:21 +00:00
/* (xr + n) * pr.z^2 mod p == pr.x, so the signature is valid. */
return 1;
}
return 0;
#endif
2015-10-26 14:54:21 +00:00
}
static int rustsecp256k1_v0_7_0_ecdsa_sig_sign(const rustsecp256k1_v0_7_0_ecmult_gen_context *ctx, rustsecp256k1_v0_7_0_scalar *sigr, rustsecp256k1_v0_7_0_scalar *sigs, const rustsecp256k1_v0_7_0_scalar *seckey, const rustsecp256k1_v0_7_0_scalar *message, const rustsecp256k1_v0_7_0_scalar *nonce, int *recid) {
2015-10-26 14:54:21 +00:00
unsigned char b[32];
rustsecp256k1_v0_7_0_gej rp;
rustsecp256k1_v0_7_0_ge r;
rustsecp256k1_v0_7_0_scalar n;
2015-10-26 14:54:21 +00:00
int overflow = 0;
int high;
2015-10-26 14:54:21 +00:00
rustsecp256k1_v0_7_0_ecmult_gen(ctx, &rp, nonce);
rustsecp256k1_v0_7_0_ge_set_gej(&r, &rp);
rustsecp256k1_v0_7_0_fe_normalize(&r.x);
rustsecp256k1_v0_7_0_fe_normalize(&r.y);
rustsecp256k1_v0_7_0_fe_get_b32(b, &r.x);
rustsecp256k1_v0_7_0_scalar_set_b32(sigr, b, &overflow);
2015-10-26 14:54:21 +00:00
if (recid) {
/* The overflow condition is cryptographically unreachable as hitting it requires finding the discrete log
* of some P where P.x >= order, and only 1 in about 2^127 points meet this criteria.
*/
*recid = (overflow << 1) | rustsecp256k1_v0_7_0_fe_is_odd(&r.y);
}
rustsecp256k1_v0_7_0_scalar_mul(&n, sigr, seckey);
rustsecp256k1_v0_7_0_scalar_add(&n, &n, message);
rustsecp256k1_v0_7_0_scalar_inverse(sigs, nonce);
rustsecp256k1_v0_7_0_scalar_mul(sigs, sigs, &n);
rustsecp256k1_v0_7_0_scalar_clear(&n);
rustsecp256k1_v0_7_0_gej_clear(&rp);
rustsecp256k1_v0_7_0_ge_clear(&r);
high = rustsecp256k1_v0_7_0_scalar_is_high(sigs);
rustsecp256k1_v0_7_0_scalar_cond_negate(sigs, high);
if (recid) {
*recid ^= high;
2015-10-26 14:54:21 +00:00
}
/* P.x = order is on the curve, so technically sig->r could end up being zero, which would be an invalid signature.
* This is cryptographically unreachable as hitting it requires finding the discrete log of P.x = N.
*/
return (int)(!rustsecp256k1_v0_7_0_scalar_is_zero(sigr)) & (int)(!rustsecp256k1_v0_7_0_scalar_is_zero(sigs));
2015-10-26 14:54:21 +00:00
}
#endif /* SECP256K1_ECDSA_IMPL_H */