| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| #ifndef BLAKE2_IMPL_H |
| #define BLAKE2_IMPL_H |
|
|
| #include <endian.h> |
| #if BYTE_ORDER == LITTLE_ENDIAN |
| # define NATIVE_LITTLE_ENDIAN 1 |
| #endif |
|
|
| #include <stdint.h> |
| #include <string.h> |
|
|
| #if !defined(__cplusplus) && (!defined(__STDC_VERSION__) || __STDC_VERSION__ < 199901L) |
| #if defined(_MSC_VER) |
| #define BLAKE2_INLINE __inline |
| #elif defined(__GNUC__) |
| #define BLAKE2_INLINE __inline__ |
| #else |
| #define BLAKE2_INLINE |
| #endif |
| #else |
| #define BLAKE2_INLINE inline |
| #endif |
|
|
| static BLAKE2_INLINE uint32_t load32( const void *src ) |
| { |
| #if defined(NATIVE_LITTLE_ENDIAN) |
| uint32_t w; |
| memcpy(&w, src, sizeof w); |
| return w; |
| #else |
| const uint8_t *p = ( const uint8_t * )src; |
| return (( uint32_t )( p[0] ) << 0) | |
| (( uint32_t )( p[1] ) << 8) | |
| (( uint32_t )( p[2] ) << 16) | |
| (( uint32_t )( p[3] ) << 24) ; |
| #endif |
| } |
|
|
| static BLAKE2_INLINE uint64_t load64( const void *src ) |
| { |
| #if defined(NATIVE_LITTLE_ENDIAN) |
| uint64_t w; |
| memcpy(&w, src, sizeof w); |
| return w; |
| #else |
| const uint8_t *p = ( const uint8_t * )src; |
| return (( uint64_t )( p[0] ) << 0) | |
| (( uint64_t )( p[1] ) << 8) | |
| (( uint64_t )( p[2] ) << 16) | |
| (( uint64_t )( p[3] ) << 24) | |
| (( uint64_t )( p[4] ) << 32) | |
| (( uint64_t )( p[5] ) << 40) | |
| (( uint64_t )( p[6] ) << 48) | |
| (( uint64_t )( p[7] ) << 56) ; |
| #endif |
| } |
|
|
| static BLAKE2_INLINE uint16_t load16( const void *src ) |
| { |
| #if defined(NATIVE_LITTLE_ENDIAN) |
| uint16_t w; |
| memcpy(&w, src, sizeof w); |
| return w; |
| #else |
| const uint8_t *p = ( const uint8_t * )src; |
| return ( uint16_t )((( uint32_t )( p[0] ) << 0) | |
| (( uint32_t )( p[1] ) << 8)); |
| #endif |
| } |
|
|
| static BLAKE2_INLINE void store16( void *dst, uint16_t w ) |
| { |
| #if defined(NATIVE_LITTLE_ENDIAN) |
| memcpy(dst, &w, sizeof w); |
| #else |
| uint8_t *p = ( uint8_t * )dst; |
| *p++ = ( uint8_t )w; w >>= 8; |
| *p++ = ( uint8_t )w; |
| #endif |
| } |
|
|
| static BLAKE2_INLINE void store32( void *dst, uint32_t w ) |
| { |
| #if defined(NATIVE_LITTLE_ENDIAN) |
| memcpy(dst, &w, sizeof w); |
| #else |
| uint8_t *p = ( uint8_t * )dst; |
| p[0] = (uint8_t)(w >> 0); |
| p[1] = (uint8_t)(w >> 8); |
| p[2] = (uint8_t)(w >> 16); |
| p[3] = (uint8_t)(w >> 24); |
| #endif |
| } |
|
|
| static BLAKE2_INLINE void store64( void *dst, uint64_t w ) |
| { |
| #if defined(NATIVE_LITTLE_ENDIAN) |
| memcpy(dst, &w, sizeof w); |
| #else |
| uint8_t *p = ( uint8_t * )dst; |
| p[0] = (uint8_t)(w >> 0); |
| p[1] = (uint8_t)(w >> 8); |
| p[2] = (uint8_t)(w >> 16); |
| p[3] = (uint8_t)(w >> 24); |
| p[4] = (uint8_t)(w >> 32); |
| p[5] = (uint8_t)(w >> 40); |
| p[6] = (uint8_t)(w >> 48); |
| p[7] = (uint8_t)(w >> 56); |
| #endif |
| } |
|
|
| static BLAKE2_INLINE uint64_t load48( const void *src ) |
| { |
| const uint8_t *p = ( const uint8_t * )src; |
| return (( uint64_t )( p[0] ) << 0) | |
| (( uint64_t )( p[1] ) << 8) | |
| (( uint64_t )( p[2] ) << 16) | |
| (( uint64_t )( p[3] ) << 24) | |
| (( uint64_t )( p[4] ) << 32) | |
| (( uint64_t )( p[5] ) << 40) ; |
| } |
|
|
| static BLAKE2_INLINE void store48( void *dst, uint64_t w ) |
| { |
| uint8_t *p = ( uint8_t * )dst; |
| p[0] = (uint8_t)(w >> 0); |
| p[1] = (uint8_t)(w >> 8); |
| p[2] = (uint8_t)(w >> 16); |
| p[3] = (uint8_t)(w >> 24); |
| p[4] = (uint8_t)(w >> 32); |
| p[5] = (uint8_t)(w >> 40); |
| } |
|
|
| static BLAKE2_INLINE uint32_t rotr32( const uint32_t w, const unsigned c ) |
| { |
| return ( w >> c ) | ( w << ( 32 - c ) ); |
| } |
|
|
| static BLAKE2_INLINE uint64_t rotr64( const uint64_t w, const unsigned c ) |
| { |
| return ( w >> c ) | ( w << ( 64 - c ) ); |
| } |
|
|
| |
| static BLAKE2_INLINE void secure_zero_memory(void *v, size_t n) |
| { |
| static void *(*const volatile memset_v)(void *, int, size_t) = &memset; |
| memset_v(v, 0, n); |
| } |
|
|
| #endif |
|
|