/* XMRig * Copyright 2010 Jeff Garzik * Copyright 2012-2014 pooler * Copyright 2014 Lucas Jones * Copyright 2014-2016 Wolf9466 * Copyright 2016 Jay D Dee * Copyright 2017-2018 XMR-Stak , * Copyright 2018 Lee Clagett * Copyright 2018-2019 SChernykh * Copyright 2016-2019 XMRig , * * This program is free software: you can redistribute it and/or modify * it under the terms of the GNU General Public License as published by * the Free Software Foundation, either version 3 of the License, or * (at your option) any later version. * * This program is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the * GNU General Public License for more details. * * You should have received a copy of the GNU General Public License * along with this program. If not, see . */ #include #include #include #include #ifndef BUILD_TEST # include "xmrig.h" #endif #include "cpu.h" #include "crypto/c_blake256.h" #include "crypto/c_groestl.h" #include "crypto/c_jh.h" #include "crypto/c_skein.h" #include "cryptonight_test.h" #include "cryptonight.h" #include "options.h" #include "persistent_memory.h" static cn_hash_fun asm_func_map[AV_MAX][VARIANT_MAX][ASM_MAX] = {}; void cryptonight_av1_v0(const uint8_t *input, size_t size, uint8_t *output, struct cryptonight_ctx **ctx); void cryptonight_av1_v1(const uint8_t *input, size_t size, uint8_t *output, struct cryptonight_ctx **ctx); void cryptonight_av1_v2(const uint8_t *input, size_t size, uint8_t *output, struct cryptonight_ctx **ctx); void cryptonight_av2_v0(const uint8_t *input, size_t size, uint8_t *output, struct cryptonight_ctx **ctx); void cryptonight_av2_v1(const uint8_t *input, size_t size, uint8_t *output, struct cryptonight_ctx **ctx); void cryptonight_av2_v2(const uint8_t *input, size_t size, uint8_t *output, struct cryptonight_ctx **ctx); void cryptonight_av3_v0(const uint8_t *input, size_t size, uint8_t *output, struct cryptonight_ctx **ctx); void cryptonight_av3_v1(const uint8_t *input, size_t size, uint8_t *output, struct cryptonight_ctx **ctx); void cryptonight_av3_v2(const uint8_t *input, size_t size, uint8_t *output, struct cryptonight_ctx **ctx); void cryptonight_av4_v0(const uint8_t *input, size_t size, uint8_t *output, struct cryptonight_ctx **ctx); void cryptonight_av4_v1(const uint8_t *input, size_t size, uint8_t *output, struct cryptonight_ctx **ctx); void cryptonight_av4_v2(const uint8_t *input, size_t size, uint8_t *output, struct cryptonight_ctx **ctx); void cryptonight_r_av1(const uint8_t *input, size_t size, uint8_t *output, struct cryptonight_ctx **ctx); void cryptonight_r_av2(const uint8_t *input, size_t size, uint8_t *output, struct cryptonight_ctx **ctx); void cryptonight_r_av3(const uint8_t *input, size_t size, uint8_t *output, struct cryptonight_ctx **ctx); void cryptonight_r_av4(const uint8_t *input, size_t size, uint8_t *output, struct cryptonight_ctx **ctx); #ifndef XMRIG_NO_AEON void cryptonight_lite_av1_v0(const uint8_t *input, size_t size, uint8_t *output, struct cryptonight_ctx **ctx); void cryptonight_lite_av1_v1(const uint8_t *input, size_t size, uint8_t *output, struct cryptonight_ctx **ctx); void cryptonight_lite_av2_v0(const uint8_t *input, size_t size, uint8_t *output, struct cryptonight_ctx **ctx); void cryptonight_lite_av2_v1(const uint8_t *input, size_t size, uint8_t *output, struct cryptonight_ctx **ctx); void cryptonight_lite_av3_v0(const uint8_t *input, size_t size, uint8_t *output, struct cryptonight_ctx **ctx); void cryptonight_lite_av3_v1(const uint8_t *input, size_t size, uint8_t *output, struct cryptonight_ctx **ctx); void cryptonight_lite_av4_v0(const uint8_t *input, size_t size, uint8_t *output, struct cryptonight_ctx **ctx); void cryptonight_lite_av4_v1(const uint8_t *input, size_t size, uint8_t *output, struct cryptonight_ctx **ctx); #endif #ifndef XMRIG_NO_ASM void cryptonight_single_hash_asm_intel(const uint8_t *input, size_t size, uint8_t *output, struct cryptonight_ctx **ctx); void cryptonight_single_hash_asm_ryzen(const uint8_t *input, size_t size, uint8_t *output, struct cryptonight_ctx **ctx); void cryptonight_single_hash_asm_bulldozer(const uint8_t *input, size_t size, uint8_t *output, struct cryptonight_ctx **ctx); void cryptonight_double_hash_asm(const uint8_t *input, size_t size, uint8_t *output, struct cryptonight_ctx **ctx); void cryptonight_r_av1_asm_intel(const uint8_t *input, size_t size, uint8_t *output, struct cryptonight_ctx **ctx); void cryptonight_r_av1_asm_bulldozer(const uint8_t *input, size_t size, uint8_t *output, struct cryptonight_ctx **ctx); void cryptonight_r_av2_asm_intel(const uint8_t *input, size_t size, uint8_t *output, struct cryptonight_ctx **ctx); void cryptonight_r_av2_asm_bulldozer(const uint8_t *input, size_t size, uint8_t *output, struct cryptonight_ctx **ctx); #endif static inline bool verify(enum Variant variant, uint8_t *output, struct cryptonight_ctx **ctx, const uint8_t *referenceValue) { cn_hash_fun func = cryptonight_hash_fn(opt_algo, opt_av, variant); if (func == NULL) { return false; } func(test_input, 76, output, ctx); return memcmp(output, referenceValue, opt_double_hash ? 64 : 32) == 0; } static inline bool verify2(enum Variant variant, uint8_t *output, struct cryptonight_ctx **ctx, const uint8_t *referenceValue) { cn_hash_fun func = cryptonight_hash_fn(opt_algo, opt_av, variant); if (func == NULL) { return false; } if (opt_double_hash) { uint8_t input[128]; for (size_t i = 0; i < (sizeof(cn_r_test_input) / sizeof(cn_r_test_input[0])); ++i) { const size_t size = cn_r_test_input[i].size; memcpy(input, cn_r_test_input[i].data, size); memcpy(input + size, cn_r_test_input[i].data, size); ctx[0]->height = ctx[1]->height = cn_r_test_input[i].height; func(input, size, output, ctx); if (memcmp(output, referenceValue + i * 32, 32) != 0 || memcmp(output + 32, referenceValue + i * 32, 32) != 0) { return false; } } } else { for (size_t i = 0; i < (sizeof(cn_r_test_input) / sizeof(cn_r_test_input[0])); ++i) { ctx[0]->height = cn_r_test_input[i].height; func(cn_r_test_input[i].data, cn_r_test_input[i].size, output, ctx); if (memcmp(output, referenceValue + i * 32, 32) != 0) { return false; } } } return true; } static bool self_test() { struct cryptonight_ctx *ctx[2]; uint8_t output[64]; const size_t count = opt_double_hash ? 2 : 1; const size_t size = opt_algo == ALGO_CRYPTONIGHT ? MEMORY : MEMORY_LITE; bool result = false; for (size_t i = 0; i < count; ++i) { ctx[i] = _mm_malloc(sizeof(struct cryptonight_ctx), 16); ctx[i]->memory = _mm_malloc(size, 16); init_cn_r(ctx[i]); } if (opt_algo == ALGO_CRYPTONIGHT) { result = verify(VARIANT_0, output, ctx, test_output_v0) && verify(VARIANT_1, output, ctx, test_output_v1) && verify(VARIANT_2, output, ctx, test_output_v2) && verify2(VARIANT_4, output, ctx, test_output_r); } # ifndef XMRIG_NO_AEON else { result = verify(VARIANT_0, output, ctx, test_output_v0_lite) && verify(VARIANT_1, output, ctx, test_output_v1_lite); } # endif for (size_t i = 0; i < count; ++i) { _mm_free(ctx[i]->memory); _mm_free(ctx[i]); } return result; } #ifndef XMRIG_NO_ASM cn_hash_fun cryptonight_hash_asm_fn(enum AlgoVariant av, enum Variant variant, enum Assembly assembly) { if (assembly == ASM_AUTO) { assembly = (enum Assembly) cpu_info.assembly; } if (assembly == ASM_NONE) { return NULL; } return asm_func_map[av][variant][assembly]; } #endif cn_hash_fun cryptonight_hash_fn(enum Algo algorithm, enum AlgoVariant av, enum Variant variant) { assert(av > AV_AUTO && av < AV_MAX); assert(variant > VARIANT_AUTO && variant < VARIANT_MAX); # ifndef XMRIG_NO_ASM if (algorithm == ALGO_CRYPTONIGHT) { cn_hash_fun fun = cryptonight_hash_asm_fn(av, variant, opt_assembly); if (fun) { return fun; } } # endif static const cn_hash_fun func_table[VARIANT_MAX * 4 * 2] = { cryptonight_av1_v0, cryptonight_av2_v0, cryptonight_av3_v0, cryptonight_av4_v0, cryptonight_av1_v1, cryptonight_av2_v1, cryptonight_av3_v1, cryptonight_av4_v1, cryptonight_av1_v2, cryptonight_av2_v2, cryptonight_av3_v2, cryptonight_av4_v2, cryptonight_r_av1, cryptonight_r_av2, cryptonight_r_av3, cryptonight_r_av4, # ifndef XMRIG_NO_AEON cryptonight_lite_av1_v0, cryptonight_lite_av2_v0, cryptonight_lite_av3_v0, cryptonight_lite_av4_v0, cryptonight_lite_av1_v1, cryptonight_lite_av2_v1, cryptonight_lite_av3_v1, cryptonight_lite_av4_v1, NULL, NULL, NULL, NULL, NULL, NULL, NULL, NULL, # else NULL, NULL, NULL, NULL, NULL, NULL, NULL, NULL, NULL, NULL, NULL, NULL, NULL, NULL, NULL, NULL, # endif }; # ifndef NDEBUG const size_t index = VARIANT_MAX * 4 * algorithm + 4 * variant + av - 1; cn_hash_fun func = func_table[index]; assert(index < sizeof(func_table) / sizeof(func_table[0])); assert(func != NULL); return func; # else return func_table[VARIANT_MAX * 4 * algorithm + 4 * variant + av - 1]; # endif } bool cryptonight_init(int av) { opt_double_hash = av == AV_DOUBLE || av == AV_DOUBLE_SOFT; # ifndef XMRIG_NO_ASM asm_func_map[AV_SINGLE][VARIANT_2][ASM_INTEL] = cryptonight_single_hash_asm_intel; asm_func_map[AV_SINGLE][VARIANT_2][ASM_RYZEN] = cryptonight_single_hash_asm_intel; asm_func_map[AV_SINGLE][VARIANT_2][ASM_BULLDOZER] = cryptonight_single_hash_asm_bulldozer; asm_func_map[AV_DOUBLE][VARIANT_2][ASM_INTEL] = cryptonight_double_hash_asm; asm_func_map[AV_DOUBLE][VARIANT_2][ASM_RYZEN] = cryptonight_double_hash_asm; asm_func_map[AV_DOUBLE][VARIANT_2][ASM_BULLDOZER] = cryptonight_double_hash_asm; asm_func_map[AV_SINGLE][VARIANT_4][ASM_INTEL] = cryptonight_r_av1_asm_intel; asm_func_map[AV_SINGLE][VARIANT_4][ASM_RYZEN] = cryptonight_r_av1_asm_intel; asm_func_map[AV_SINGLE][VARIANT_4][ASM_BULLDOZER] = cryptonight_r_av1_asm_bulldozer; asm_func_map[AV_DOUBLE][VARIANT_4][ASM_INTEL] = cryptonight_r_av2_asm_intel; asm_func_map[AV_DOUBLE][VARIANT_4][ASM_RYZEN] = cryptonight_r_av2_asm_intel; asm_func_map[AV_DOUBLE][VARIANT_4][ASM_BULLDOZER] = cryptonight_r_av2_asm_bulldozer; # endif return self_test(); } static inline void do_blake_hash(const void* input, size_t len, char* output) { blake256_hash((uint8_t*)output, input, len); } static inline void do_groestl_hash(const void* input, size_t len, char* output) { groestl(input, len * 8, (uint8_t*)output); } static inline void do_jh_hash(const void* input, size_t len, char* output) { jh_hash(32 * 8, input, 8 * len, (uint8_t*)output); } static inline void do_skein_hash(const void* input, size_t len, char* output) { skein_hash(8 * 32, input, 8 * len, (uint8_t*)output); } void (* const extra_hashes[4])(const void *, size_t, char *) = {do_blake_hash, do_groestl_hash, do_jh_hash, do_skein_hash}; static inline enum Variant cryptonight_variant(uint8_t version) { if (opt_variant != VARIANT_AUTO) { return opt_variant; } if (opt_algo == ALGO_CRYPTONIGHT_LITE) { return VARIANT_1; } if (version >= 10) { return VARIANT_4; } if (version >= 8) { return VARIANT_2; } return version == 7 ? VARIANT_1 : VARIANT_0; } #ifndef BUILD_TEST int scanhash_cryptonight(int thr_id, uint32_t *hash, uint8_t *restrict blob, size_t blob_size, uint32_t target, uint32_t max_nonce, unsigned long *restrict hashes_done, struct cryptonight_ctx **restrict ctx) { uint32_t *nonceptr = (uint32_t*) (((char*) blob) + 39); enum Variant variant = cryptonight_variant(blob[0]); do { cryptonight_hash_fn(opt_algo, opt_av, variant)(blob, blob_size, (uint8_t *) hash, ctx); (*hashes_done)++; if (unlikely(hash[7] < target)) { return 1; } (*nonceptr)++; } while (likely(((*nonceptr) < max_nonce && !work_restart[thr_id].restart))); return 0; } int scanhash_cryptonight_double(int thr_id, uint32_t *hash, uint8_t *restrict blob, size_t blob_size, uint32_t target, uint32_t max_nonce, unsigned long *restrict hashes_done, struct cryptonight_ctx **restrict ctx) { int rc = 0; uint32_t *nonceptr0 = (uint32_t*) (((char*) blob) + 39); uint32_t *nonceptr1 = (uint32_t*) (((char*) blob) + 39 + blob_size); enum Variant variant = cryptonight_variant(blob[0]); do { cryptonight_hash_fn(opt_algo, opt_av, variant)(blob, blob_size, (uint8_t *) hash, ctx); (*hashes_done) += 2; if (unlikely(hash[7] < target)) { return rc |= 1; } if (unlikely(hash[15] < target)) { return rc |= 2; } if (rc) { break; } (*nonceptr0)++; (*nonceptr1)++; } while (likely(((*nonceptr0) < max_nonce && !work_restart[thr_id].restart))); return rc; } #endif