@@ -42,18 +42,18 @@ SECP256K1_INLINE static void secp256k1_scalar_set_int(secp256k1_scalar *r, unsig
42
42
r -> d [2 ] = 0 ;
43
43
r -> d [3 ] = 0 ;
44
44
45
- secp256k1_scalar_verify (r );
45
+ SECP256K1_SCALAR_VERIFY (r );
46
46
}
47
47
48
48
SECP256K1_INLINE static unsigned int secp256k1_scalar_get_bits (const secp256k1_scalar * a , unsigned int offset , unsigned int count ) {
49
- secp256k1_scalar_verify (a );
49
+ SECP256K1_SCALAR_VERIFY (a );
50
50
VERIFY_CHECK ((offset + count - 1 ) >> 6 == offset >> 6 );
51
51
52
52
return (a -> d [offset >> 6 ] >> (offset & 0x3F )) & ((((uint64_t )1 ) << count ) - 1 );
53
53
}
54
54
55
55
SECP256K1_INLINE static unsigned int secp256k1_scalar_get_bits_var (const secp256k1_scalar * a , unsigned int offset , unsigned int count ) {
56
- secp256k1_scalar_verify (a );
56
+ SECP256K1_SCALAR_VERIFY (a );
57
57
VERIFY_CHECK (count < 32 );
58
58
VERIFY_CHECK (offset + count <= 256 );
59
59
@@ -93,15 +93,15 @@ SECP256K1_INLINE static int secp256k1_scalar_reduce(secp256k1_scalar *r, unsigne
93
93
secp256k1_u128_accum_u64 (& t , r -> d [3 ]);
94
94
r -> d [3 ] = secp256k1_u128_to_u64 (& t );
95
95
96
- secp256k1_scalar_verify (r );
96
+ SECP256K1_SCALAR_VERIFY (r );
97
97
return overflow ;
98
98
}
99
99
100
100
static int secp256k1_scalar_add (secp256k1_scalar * r , const secp256k1_scalar * a , const secp256k1_scalar * b ) {
101
101
int overflow ;
102
102
secp256k1_uint128 t ;
103
- secp256k1_scalar_verify (a );
104
- secp256k1_scalar_verify (b );
103
+ SECP256K1_SCALAR_VERIFY (a );
104
+ SECP256K1_SCALAR_VERIFY (b );
105
105
106
106
secp256k1_u128_from_u64 (& t , a -> d [0 ]);
107
107
secp256k1_u128_accum_u64 (& t , b -> d [0 ]);
@@ -119,14 +119,14 @@ static int secp256k1_scalar_add(secp256k1_scalar *r, const secp256k1_scalar *a,
119
119
VERIFY_CHECK (overflow == 0 || overflow == 1 );
120
120
secp256k1_scalar_reduce (r , overflow );
121
121
122
- secp256k1_scalar_verify (r );
122
+ SECP256K1_SCALAR_VERIFY (r );
123
123
return overflow ;
124
124
}
125
125
126
126
static void secp256k1_scalar_cadd_bit (secp256k1_scalar * r , unsigned int bit , int flag ) {
127
127
secp256k1_uint128 t ;
128
128
volatile int vflag = flag ;
129
- secp256k1_scalar_verify (r );
129
+ SECP256K1_SCALAR_VERIFY (r );
130
130
VERIFY_CHECK (bit < 256 );
131
131
132
132
bit += ((uint32_t ) vflag - 1 ) & 0x100 ; /* forcing (bit >> 6) > 3 makes this a noop */
@@ -143,7 +143,7 @@ static void secp256k1_scalar_cadd_bit(secp256k1_scalar *r, unsigned int bit, int
143
143
secp256k1_u128_accum_u64 (& t , ((uint64_t )((bit >> 6 ) == 3 )) << (bit & 0x3F ));
144
144
r -> d [3 ] = secp256k1_u128_to_u64 (& t );
145
145
146
- secp256k1_scalar_verify (r );
146
+ SECP256K1_SCALAR_VERIFY (r );
147
147
VERIFY_CHECK (secp256k1_u128_hi_u64 (& t ) == 0 );
148
148
}
149
149
@@ -158,11 +158,11 @@ static void secp256k1_scalar_set_b32(secp256k1_scalar *r, const unsigned char *b
158
158
* overflow = over ;
159
159
}
160
160
161
- secp256k1_scalar_verify (r );
161
+ SECP256K1_SCALAR_VERIFY (r );
162
162
}
163
163
164
164
static void secp256k1_scalar_get_b32 (unsigned char * bin , const secp256k1_scalar * a ) {
165
- secp256k1_scalar_verify (a );
165
+ SECP256K1_SCALAR_VERIFY (a );
166
166
167
167
secp256k1_write_be64 (& bin [0 ], a -> d [3 ]);
168
168
secp256k1_write_be64 (& bin [8 ], a -> d [2 ]);
@@ -171,15 +171,15 @@ static void secp256k1_scalar_get_b32(unsigned char *bin, const secp256k1_scalar*
171
171
}
172
172
173
173
SECP256K1_INLINE static int secp256k1_scalar_is_zero (const secp256k1_scalar * a ) {
174
- secp256k1_scalar_verify (a );
174
+ SECP256K1_SCALAR_VERIFY (a );
175
175
176
176
return (a -> d [0 ] | a -> d [1 ] | a -> d [2 ] | a -> d [3 ]) == 0 ;
177
177
}
178
178
179
179
static void secp256k1_scalar_negate (secp256k1_scalar * r , const secp256k1_scalar * a ) {
180
180
uint64_t nonzero = 0xFFFFFFFFFFFFFFFFULL * (secp256k1_scalar_is_zero (a ) == 0 );
181
181
secp256k1_uint128 t ;
182
- secp256k1_scalar_verify (a );
182
+ SECP256K1_SCALAR_VERIFY (a );
183
183
184
184
secp256k1_u128_from_u64 (& t , ~a -> d [0 ]);
185
185
secp256k1_u128_accum_u64 (& t , SECP256K1_N_0 + 1 );
@@ -194,7 +194,7 @@ static void secp256k1_scalar_negate(secp256k1_scalar *r, const secp256k1_scalar
194
194
secp256k1_u128_accum_u64 (& t , SECP256K1_N_3 );
195
195
r -> d [3 ] = secp256k1_u128_to_u64 (& t ) & nonzero ;
196
196
197
- secp256k1_scalar_verify (r );
197
+ SECP256K1_SCALAR_VERIFY (r );
198
198
}
199
199
200
200
static void secp256k1_scalar_half (secp256k1_scalar * r , const secp256k1_scalar * a ) {
@@ -214,7 +214,7 @@ static void secp256k1_scalar_half(secp256k1_scalar *r, const secp256k1_scalar *a
214
214
*/
215
215
uint64_t mask = - (uint64_t )(a -> d [0 ] & 1U );
216
216
secp256k1_uint128 t ;
217
- secp256k1_scalar_verify (a );
217
+ SECP256K1_SCALAR_VERIFY (a );
218
218
219
219
secp256k1_u128_from_u64 (& t , (a -> d [0 ] >> 1 ) | (a -> d [1 ] << 63 ));
220
220
secp256k1_u128_accum_u64 (& t , (SECP256K1_N_H_0 + 1U ) & mask );
@@ -234,20 +234,20 @@ static void secp256k1_scalar_half(secp256k1_scalar *r, const secp256k1_scalar *a
234
234
secp256k1_u128_rshift (& t , 64 );
235
235
VERIFY_CHECK (secp256k1_u128_to_u64 (& t ) == 0 );
236
236
237
- secp256k1_scalar_verify (r );
237
+ SECP256K1_SCALAR_VERIFY (r );
238
238
#endif
239
239
}
240
240
241
241
SECP256K1_INLINE static int secp256k1_scalar_is_one (const secp256k1_scalar * a ) {
242
- secp256k1_scalar_verify (a );
242
+ SECP256K1_SCALAR_VERIFY (a );
243
243
244
244
return ((a -> d [0 ] ^ 1 ) | a -> d [1 ] | a -> d [2 ] | a -> d [3 ]) == 0 ;
245
245
}
246
246
247
247
static int secp256k1_scalar_is_high (const secp256k1_scalar * a ) {
248
248
int yes = 0 ;
249
249
int no = 0 ;
250
- secp256k1_scalar_verify (a );
250
+ SECP256K1_SCALAR_VERIFY (a );
251
251
252
252
no |= (a -> d [3 ] < SECP256K1_N_H_3 );
253
253
yes |= (a -> d [3 ] > SECP256K1_N_H_3 ) & ~no ;
@@ -265,7 +265,7 @@ static int secp256k1_scalar_cond_negate(secp256k1_scalar *r, int flag) {
265
265
uint64_t mask = - vflag ;
266
266
uint64_t nonzero = (secp256k1_scalar_is_zero (r ) != 0 ) - 1 ;
267
267
secp256k1_uint128 t ;
268
- secp256k1_scalar_verify (r );
268
+ SECP256K1_SCALAR_VERIFY (r );
269
269
270
270
secp256k1_u128_from_u64 (& t , r -> d [0 ] ^ mask );
271
271
secp256k1_u128_accum_u64 (& t , (SECP256K1_N_0 + 1 ) & mask );
@@ -280,7 +280,7 @@ static int secp256k1_scalar_cond_negate(secp256k1_scalar *r, int flag) {
280
280
secp256k1_u128_accum_u64 (& t , SECP256K1_N_3 & mask );
281
281
r -> d [3 ] = secp256k1_u128_to_u64 (& t ) & nonzero ;
282
282
283
- secp256k1_scalar_verify (r );
283
+ SECP256K1_SCALAR_VERIFY (r );
284
284
return 2 * (mask == 0 ) - 1 ;
285
285
}
286
286
@@ -839,17 +839,17 @@ static void secp256k1_scalar_mul_512(uint64_t l[8], const secp256k1_scalar *a, c
839
839
840
840
static void secp256k1_scalar_mul (secp256k1_scalar * r , const secp256k1_scalar * a , const secp256k1_scalar * b ) {
841
841
uint64_t l [8 ];
842
- secp256k1_scalar_verify (a );
843
- secp256k1_scalar_verify (b );
842
+ SECP256K1_SCALAR_VERIFY (a );
843
+ SECP256K1_SCALAR_VERIFY (b );
844
844
845
845
secp256k1_scalar_mul_512 (l , a , b );
846
846
secp256k1_scalar_reduce_512 (r , l );
847
847
848
- secp256k1_scalar_verify (r );
848
+ SECP256K1_SCALAR_VERIFY (r );
849
849
}
850
850
851
851
static void secp256k1_scalar_split_128 (secp256k1_scalar * r1 , secp256k1_scalar * r2 , const secp256k1_scalar * k ) {
852
- secp256k1_scalar_verify (k );
852
+ SECP256K1_SCALAR_VERIFY (k );
853
853
854
854
r1 -> d [0 ] = k -> d [0 ];
855
855
r1 -> d [1 ] = k -> d [1 ];
@@ -860,13 +860,13 @@ static void secp256k1_scalar_split_128(secp256k1_scalar *r1, secp256k1_scalar *r
860
860
r2 -> d [2 ] = 0 ;
861
861
r2 -> d [3 ] = 0 ;
862
862
863
- secp256k1_scalar_verify (r1 );
864
- secp256k1_scalar_verify (r2 );
863
+ SECP256K1_SCALAR_VERIFY (r1 );
864
+ SECP256K1_SCALAR_VERIFY (r2 );
865
865
}
866
866
867
867
SECP256K1_INLINE static int secp256k1_scalar_eq (const secp256k1_scalar * a , const secp256k1_scalar * b ) {
868
- secp256k1_scalar_verify (a );
869
- secp256k1_scalar_verify (b );
868
+ SECP256K1_SCALAR_VERIFY (a );
869
+ SECP256K1_SCALAR_VERIFY (b );
870
870
871
871
return ((a -> d [0 ] ^ b -> d [0 ]) | (a -> d [1 ] ^ b -> d [1 ]) | (a -> d [2 ] ^ b -> d [2 ]) | (a -> d [3 ] ^ b -> d [3 ])) == 0 ;
872
872
}
@@ -876,8 +876,8 @@ SECP256K1_INLINE static void secp256k1_scalar_mul_shift_var(secp256k1_scalar *r,
876
876
unsigned int shiftlimbs ;
877
877
unsigned int shiftlow ;
878
878
unsigned int shifthigh ;
879
- secp256k1_scalar_verify (a );
880
- secp256k1_scalar_verify (b );
879
+ SECP256K1_SCALAR_VERIFY (a );
880
+ SECP256K1_SCALAR_VERIFY (b );
881
881
VERIFY_CHECK (shift >= 256 );
882
882
883
883
secp256k1_scalar_mul_512 (l , a , b );
@@ -890,13 +890,13 @@ SECP256K1_INLINE static void secp256k1_scalar_mul_shift_var(secp256k1_scalar *r,
890
890
r -> d [3 ] = shift < 320 ? (l [3 + shiftlimbs ] >> shiftlow ) : 0 ;
891
891
secp256k1_scalar_cadd_bit (r , 0 , (l [(shift - 1 ) >> 6 ] >> ((shift - 1 ) & 0x3f )) & 1 );
892
892
893
- secp256k1_scalar_verify (r );
893
+ SECP256K1_SCALAR_VERIFY (r );
894
894
}
895
895
896
896
static SECP256K1_INLINE void secp256k1_scalar_cmov (secp256k1_scalar * r , const secp256k1_scalar * a , int flag ) {
897
897
uint64_t mask0 , mask1 ;
898
898
volatile int vflag = flag ;
899
- secp256k1_scalar_verify (a );
899
+ SECP256K1_SCALAR_VERIFY (a );
900
900
SECP256K1_CHECKMEM_CHECK_VERIFY (r -> d , sizeof (r -> d ));
901
901
902
902
mask0 = vflag + ~((uint64_t )0 );
@@ -906,7 +906,7 @@ static SECP256K1_INLINE void secp256k1_scalar_cmov(secp256k1_scalar *r, const se
906
906
r -> d [2 ] = (r -> d [2 ] & mask0 ) | (a -> d [2 ] & mask1 );
907
907
r -> d [3 ] = (r -> d [3 ] & mask0 ) | (a -> d [3 ] & mask1 );
908
908
909
- secp256k1_scalar_verify (r );
909
+ SECP256K1_SCALAR_VERIFY (r );
910
910
}
911
911
912
912
static void secp256k1_scalar_from_signed62 (secp256k1_scalar * r , const secp256k1_modinv64_signed62 * a ) {
@@ -926,13 +926,13 @@ static void secp256k1_scalar_from_signed62(secp256k1_scalar *r, const secp256k1_
926
926
r -> d [2 ] = a2 >> 4 | a3 << 58 ;
927
927
r -> d [3 ] = a3 >> 6 | a4 << 56 ;
928
928
929
- secp256k1_scalar_verify (r );
929
+ SECP256K1_SCALAR_VERIFY (r );
930
930
}
931
931
932
932
static void secp256k1_scalar_to_signed62 (secp256k1_modinv64_signed62 * r , const secp256k1_scalar * a ) {
933
933
const uint64_t M62 = UINT64_MAX >> 2 ;
934
934
const uint64_t a0 = a -> d [0 ], a1 = a -> d [1 ], a2 = a -> d [2 ], a3 = a -> d [3 ];
935
- secp256k1_scalar_verify (a );
935
+ SECP256K1_SCALAR_VERIFY (a );
936
936
937
937
r -> v [0 ] = a0 & M62 ;
938
938
r -> v [1 ] = (a0 >> 62 | a1 << 2 ) & M62 ;
@@ -951,13 +951,13 @@ static void secp256k1_scalar_inverse(secp256k1_scalar *r, const secp256k1_scalar
951
951
#ifdef VERIFY
952
952
int zero_in = secp256k1_scalar_is_zero (x );
953
953
#endif
954
- secp256k1_scalar_verify (x );
954
+ SECP256K1_SCALAR_VERIFY (x );
955
955
956
956
secp256k1_scalar_to_signed62 (& s , x );
957
957
secp256k1_modinv64 (& s , & secp256k1_const_modinfo_scalar );
958
958
secp256k1_scalar_from_signed62 (r , & s );
959
959
960
- secp256k1_scalar_verify (r );
960
+ SECP256K1_SCALAR_VERIFY (r );
961
961
VERIFY_CHECK (secp256k1_scalar_is_zero (r ) == zero_in );
962
962
}
963
963
@@ -966,18 +966,18 @@ static void secp256k1_scalar_inverse_var(secp256k1_scalar *r, const secp256k1_sc
966
966
#ifdef VERIFY
967
967
int zero_in = secp256k1_scalar_is_zero (x );
968
968
#endif
969
- secp256k1_scalar_verify (x );
969
+ SECP256K1_SCALAR_VERIFY (x );
970
970
971
971
secp256k1_scalar_to_signed62 (& s , x );
972
972
secp256k1_modinv64_var (& s , & secp256k1_const_modinfo_scalar );
973
973
secp256k1_scalar_from_signed62 (r , & s );
974
974
975
- secp256k1_scalar_verify (r );
975
+ SECP256K1_SCALAR_VERIFY (r );
976
976
VERIFY_CHECK (secp256k1_scalar_is_zero (r ) == zero_in );
977
977
}
978
978
979
979
SECP256K1_INLINE static int secp256k1_scalar_is_even (const secp256k1_scalar * a ) {
980
- secp256k1_scalar_verify (a );
980
+ SECP256K1_SCALAR_VERIFY (a );
981
981
982
982
return !(a -> d [0 ] & 1 );
983
983
}
0 commit comments