diff --git a/src/confid.rs b/src/confid.rs index d6c9e40..7812d38 100644 --- a/src/confid.rs +++ b/src/confid.rs @@ -228,8 +228,8 @@ unsafe fn find_divisor_v(mut d: *mut TDivisor) -> i32 { f2[i as usize] = f[i as usize]; i += 1; } - let u0: u64 = (*d).u[0_i32 as usize] as u64; - let u1: u64 = (*d).u[1_i32 as usize] as u64; + let u0: u64 = (*d).u[0_i32 as usize]; + let u1: u64 = (*d).u[1_i32 as usize]; let mut j: i32 = 4_i32; loop { let fresh0 = j; @@ -492,18 +492,18 @@ unsafe fn polynomial_xgcd( } unsafe fn u2poly(src: *const TDivisor, polyu: *mut u64, polyv: *mut u64) -> i32 { - if (*src).u[1_i32 as usize] as u64 != BAD { - *polyu.offset(0_i32 as isize) = (*src).u[0_i32 as usize] as u64; - *polyu.offset(1_i32 as isize) = (*src).u[1_i32 as usize] as u64; + if (*src).u[1_i32 as usize] != BAD { + *polyu.offset(0_i32 as isize) = (*src).u[0_i32 as usize]; + *polyu.offset(1_i32 as isize) = (*src).u[1_i32 as usize]; *polyu.offset(2_i32 as isize) = 1_i32 as u64; - *polyv.offset(0_i32 as isize) = (*src).v[0_i32 as usize] as u64; - *polyv.offset(1_i32 as isize) = (*src).v[1_i32 as usize] as u64; + *polyv.offset(0_i32 as isize) = (*src).v[0_i32 as usize]; + *polyv.offset(1_i32 as isize) = (*src).v[1_i32 as usize]; return 2_i32; } - if (*src).u[0_i32 as usize] as u64 != BAD { - *polyu.offset(0_i32 as isize) = (*src).u[0_i32 as usize] as u64; + if (*src).u[0_i32 as usize] != BAD { + *polyu.offset(0_i32 as isize) = (*src).u[0_i32 as usize]; *polyu.offset(1_i32 as isize) = 1_i32 as u64; - *polyv.offset(0_i32 as isize) = (*src).v[0_i32 as usize] as u64; + *polyv.offset(0_i32 as isize) = (*src).v[0_i32 as usize]; *polyv.offset(1_i32 as isize) = 0_i32 as u64; return 1_i32; } @@ -752,24 +752,24 @@ unsafe fn divisor_add(src1: *const TDivisor, src2: *const TDivisor, mut dst: *mu if udeg == 2_i32 { (*dst).u[0_i32 as usize] = u[0_i32 as usize]; (*dst).u[1_i32 as usize] = u[1_i32 as usize]; - (*dst).v[0_i32 as usize] = (if vdeg >= 0_i32 { + (*dst).v[0_i32 as usize] = if vdeg >= 0_i32 { v[0_i32 as usize] } else { 0_i32 as u64 - }); - (*dst).v[1_i32 as usize] = (if vdeg >= 1_i32 { + }; + (*dst).v[1_i32 as usize] = if vdeg >= 1_i32 { v[1_i32 as usize] } else { 0_i32 as u64 - }); + }; } else if udeg == 1_i32 { (*dst).u[0_i32 as usize] = u[0_i32 as usize]; (*dst).u[1_i32 as usize] = BAD; - (*dst).v[0_i32 as usize] = (if vdeg >= 0_i32 { + (*dst).v[0_i32 as usize] = if vdeg >= 0_i32 { v[0_i32 as usize] } else { 0_i32 as u64 - }); + }; (*dst).v[1_i32 as usize] = BAD; } else { (*dst).u[0_i32 as usize] = BAD; @@ -1201,17 +1201,17 @@ unsafe fn Generate(installation_id_str: *const i8, confirmation_id: *mut i8) -> encoded_hi: 0, }, }; - if d_0.u[0_i32 as usize] as u64 == BAD { + if d_0.u[0_i32 as usize] == BAD { // we can not get the zero divisor, actually... e.c2rust_unnamed.encoded_lo = umul128( MOD.wrapping_add(2_i32 as u64), MOD, &mut e.c2rust_unnamed.encoded_hi, ); - } else if d_0.u[1_i32 as usize] as u64 == BAD { + } else if d_0.u[1_i32 as usize] == BAD { e.c2rust_unnamed.encoded_lo = umul128( MOD.wrapping_add(1_i32 as u64), - d_0.u[0_i32 as usize] as u64, + d_0.u[0_i32 as usize], &mut e.c2rust_unnamed.encoded_hi, ); e.c2rust_unnamed.encoded_lo = e.c2rust_unnamed.encoded_lo.wrapping_add(MOD); @@ -1221,12 +1221,12 @@ unsafe fn Generate(installation_id_str: *const i8, confirmation_id: *mut i8) -> .wrapping_add((e.c2rust_unnamed.encoded_lo < MOD) as i32 as u64); } else { let x1_0: u64 = (if d_0.u[1_i32 as usize] as i32 % 2_i32 != 0 { - (d_0.u[1_i32 as usize] as u64).wrapping_add(MOD) + d_0.u[1_i32 as usize].wrapping_add(MOD) } else { - d_0.u[1_i32 as usize] as u64 + d_0.u[1_i32 as usize] }) .wrapping_div(2_i32 as u64); - let x2sqr: u64 = residue_sub(residue_mul(x1_0, x1_0), d_0.u[0_i32 as usize] as u64); + let x2sqr: u64 = residue_sub(residue_mul(x1_0, x1_0), d_0.u[0_i32 as usize]); let mut x2_0: u64 = residue_sqrt(x2sqr); if x2_0 == BAD { x2_0 = residue_sqrt(residue_mul(x2sqr, residue_inv(43_i32 as u64))); @@ -1244,13 +1244,13 @@ unsafe fn Generate(installation_id_str: *const i8, confirmation_id: *mut i8) -> // points (-x1+x2, v(-x1+x2)) and (-x1-x2, v(-x1-x2)) let mut x1a: u64 = residue_sub(x1_0, x2_0); let y1: u64 = residue_sub( - d_0.v[0_i32 as usize] as u64, - residue_mul(d_0.v[1_i32 as usize] as u64, x1a), + d_0.v[0_i32 as usize], + residue_mul(d_0.v[1_i32 as usize], x1a), ); let mut x2a: u64 = residue_add(x1_0, x2_0); let y2: u64 = residue_sub( - d_0.v[0_i32 as usize] as u64, - residue_mul(d_0.v[1_i32 as usize] as u64, x2a), + d_0.v[0_i32 as usize], + residue_mul(d_0.v[1_i32 as usize], x2a), ); if x1a > x2a { std::mem::swap(&mut x1a, &mut x2a);