// Copyright (c) 2019-2021 Alexander Medvednikov. All rights reserved. // Use of this source code is governed by an MIT license // that can be found in the LICENSE file. // This is the generic version with no architecture optimizations. // In its own file so that an architecture // optimized verision can be substituted module md5 import math.bits import encoding.binary fn block_generic(mut dig Digest, p []byte) { // load state mut a := dig.s[0] mut b := dig.s[1] mut c := dig.s[2] mut d := dig.s[3] for i := 0; i <= p.len-block_size; i += block_size { mut q := p[i..] q = q[..block_size] // save current state aa := a bb := b cc := c dd := d // load input block x0 := binary.little_endian_u32(q[4*0x0..]) x1 := binary.little_endian_u32(q[4*0x1..]) x2 := binary.little_endian_u32(q[4*0x2..]) x3 := binary.little_endian_u32(q[4*0x3..]) x4 := binary.little_endian_u32(q[4*0x4..]) x5 := binary.little_endian_u32(q[4*0x5..]) x6 := binary.little_endian_u32(q[4*0x6..]) x7 := binary.little_endian_u32(q[4*0x7..]) x8 := binary.little_endian_u32(q[4*0x8..]) x9 := binary.little_endian_u32(q[4*0x9..]) xa := binary.little_endian_u32(q[4*0xa..]) xb := binary.little_endian_u32(q[4*0xb..]) xc := binary.little_endian_u32(q[4*0xc..]) xd := binary.little_endian_u32(q[4*0xd..]) xe := binary.little_endian_u32(q[4*0xe..]) xf := binary.little_endian_u32(q[4*0xf..]) // round 1 a = b + bits.rotate_left_32((((c^d)&b)^d)+a+x0+u32(0xd76aa478), 7) d = a + bits.rotate_left_32((((b^c)&a)^c)+d+x1+u32(0xe8c7b756), 12) c = d + bits.rotate_left_32((((a^b)&d)^b)+c+x2+u32(0x242070db), 17) b = c + bits.rotate_left_32((((d^a)&c)^a)+b+x3+u32(0xc1bdceee), 22) a = b + bits.rotate_left_32((((c^d)&b)^d)+a+x4+u32(0xf57c0faf), 7) d = a + bits.rotate_left_32((((b^c)&a)^c)+d+x5+u32(0x4787c62a), 12) c = d + bits.rotate_left_32((((a^b)&d)^b)+c+x6+u32(0xa8304613), 17) b = c + bits.rotate_left_32((((d^a)&c)^a)+b+x7+u32(0xfd469501), 22) a = b + bits.rotate_left_32((((c^d)&b)^d)+a+x8+u32(0x698098d8), 7) d = a + bits.rotate_left_32((((b^c)&a)^c)+d+x9+u32(0x8b44f7af), 12) c = d + bits.rotate_left_32((((a^b)&d)^b)+c+xa+u32(0xffff5bb1), 17) b = c + bits.rotate_left_32((((d^a)&c)^a)+b+xb+u32(0x895cd7be), 22) a = b + bits.rotate_left_32((((c^d)&b)^d)+a+xc+u32(0x6b901122), 7) d = a + bits.rotate_left_32((((b^c)&a)^c)+d+xd+u32(0xfd987193), 12) c = d + bits.rotate_left_32((((a^b)&d)^b)+c+xe+u32(0xa679438e), 17) b = c + bits.rotate_left_32((((d^a)&c)^a)+b+xf+u32(0x49b40821), 22) // round 2 a = b + bits.rotate_left_32((((b^c)&d)^c)+a+x1+u32(0xf61e2562), 5) d = a + bits.rotate_left_32((((a^b)&c)^b)+d+x6+u32(0xc040b340), 9) c = d + bits.rotate_left_32((((d^a)&b)^a)+c+xb+u32(0x265e5a51), 14) b = c + bits.rotate_left_32((((c^d)&a)^d)+b+x0+u32(0xe9b6c7aa), 20) a = b + bits.rotate_left_32((((b^c)&d)^c)+a+x5+u32(0xd62f105d), 5) d = a + bits.rotate_left_32((((a^b)&c)^b)+d+xa+u32(0x02441453), 9) c = d + bits.rotate_left_32((((d^a)&b)^a)+c+xf+u32(0xd8a1e681), 14) b = c + bits.rotate_left_32((((c^d)&a)^d)+b+x4+u32(0xe7d3fbc8), 20) a = b + bits.rotate_left_32((((b^c)&d)^c)+a+x9+u32(0x21e1cde6), 5) d = a + bits.rotate_left_32((((a^b)&c)^b)+d+xe+u32(0xc33707d6), 9) c = d + bits.rotate_left_32((((d^a)&b)^a)+c+x3+u32(0xf4d50d87), 14) b = c + bits.rotate_left_32((((c^d)&a)^d)+b+x8+u32(0x455a14ed), 20) a = b + bits.rotate_left_32((((b^c)&d)^c)+a+xd+u32(0xa9e3e905), 5) d = a + bits.rotate_left_32((((a^b)&c)^b)+d+x2+u32(0xfcefa3f8), 9) c = d + bits.rotate_left_32((((d^a)&b)^a)+c+x7+u32(0x676f02d9), 14) b = c + bits.rotate_left_32((((c^d)&a)^d)+b+xc+u32(0x8d2a4c8a), 20) // round 3 a = b + bits.rotate_left_32((b^c^d)+a+x5+u32(0xfffa3942), 4) d = a + bits.rotate_left_32((a^b^c)+d+x8+u32(0x8771f681), 11) c = d + bits.rotate_left_32((d^a^b)+c+xb+u32(0x6d9d6122), 16) b = c + bits.rotate_left_32((c^d^a)+b+xe+u32(0xfde5380c), 23) a = b + bits.rotate_left_32((b^c^d)+a+x1+u32(0xa4beea44), 4) d = a + bits.rotate_left_32((a^b^c)+d+x4+u32(0x4bdecfa9), 11) c = d + bits.rotate_left_32((d^a^b)+c+x7+u32(0xf6bb4b60), 16) b = c + bits.rotate_left_32((c^d^a)+b+xa+u32(0xbebfbc70), 23) a = b + bits.rotate_left_32((b^c^d)+a+xd+u32(0x289b7ec6), 4) d = a + bits.rotate_left_32((a^b^c)+d+x0+u32(0xeaa127fa), 11) c = d + bits.rotate_left_32((d^a^b)+c+x3+u32(0xd4ef3085), 16) b = c + bits.rotate_left_32((c^d^a)+b+x6+u32(0x04881d05), 23) a = b + bits.rotate_left_32((b^c^d)+a+x9+u32(0xd9d4d039), 4) d = a + bits.rotate_left_32((a^b^c)+d+xc+u32(0xe6db99e5), 11) c = d + bits.rotate_left_32((d^a^b)+c+xf+u32(0x1fa27cf8), 16) b = c + bits.rotate_left_32((c^d^a)+b+x2+u32(0xc4ac5665), 23) // round 4 a = b + bits.rotate_left_32((c^(b|~d))+a+x0+u32(0xf4292244), 6) d = a + bits.rotate_left_32((b^(a|~c))+d+x7+u32(0x432aff97), 10) c = d + bits.rotate_left_32((a^(d|~b))+c+xe+u32(0xab9423a7), 15) b = c + bits.rotate_left_32((d^(c|~a))+b+x5+u32(0xfc93a039), 21) a = b + bits.rotate_left_32((c^(b|~d))+a+xc+u32(0x655b59c3), 6) d = a + bits.rotate_left_32((b^(a|~c))+d+x3+u32(0x8f0ccc92), 10) c = d + bits.rotate_left_32((a^(d|~b))+c+xa+u32(0xffeff47d), 15) b = c + bits.rotate_left_32((d^(c|~a))+b+x1+u32(0x85845dd1), 21) a = b + bits.rotate_left_32((c^(b|~d))+a+x8+u32(0x6fa87e4f), 6) d = a + bits.rotate_left_32((b^(a|~c))+d+xf+u32(0xfe2ce6e0), 10) c = d + bits.rotate_left_32((a^(d|~b))+c+x6+u32(0xa3014314), 15) b = c + bits.rotate_left_32((d^(c|~a))+b+xd+u32(0x4e0811a1), 21) a = b + bits.rotate_left_32((c^(b|~d))+a+x4+u32(0xf7537e82), 6) d = a + bits.rotate_left_32((b^(a|~c))+d+xb+u32(0xbd3af235), 10) c = d + bits.rotate_left_32((a^(d|~b))+c+x2+u32(0x2ad7d2bb), 15) b = c + bits.rotate_left_32((d^(c|~a))+b+x9+u32(0xeb86d391), 21) // add saved state a += aa b += bb c += cc d += dd } // save state dig.s[0] = a dig.s[1] = b dig.s[2] = c dig.s[3] = d }