mirror of
https://github.com/SChernykh/p2pool.git
synced 2024-12-22 11:29:23 +00:00
Added more integrity checks
This commit is contained in:
parent
c75bc83fc1
commit
1554a44f5c
2 changed files with 46 additions and 1 deletions
|
@ -911,7 +911,21 @@ uint32_t BlockTemplate::get_hashing_blobs(uint32_t extra_nonce_start, uint32_t c
|
|||
|
||||
for (uint32_t i = 0; i < count; ++i) {
|
||||
uint8_t blob[128];
|
||||
blob_size = get_hashing_blob_nolock(extra_nonce_start + i, blob);
|
||||
const uint32_t n = get_hashing_blob_nolock(extra_nonce_start + i, blob);
|
||||
|
||||
if (n > sizeof(blob)) {
|
||||
LOGERR(1, "internal error: get_hashing_blob_nolock returned too large blob size " << n << ", expected <= " << sizeof(blob));
|
||||
}
|
||||
else if (n < 76) {
|
||||
LOGERR(1, "internal error: get_hashing_blob_nolock returned too little blob size " << n << ", expected >= 76");
|
||||
}
|
||||
|
||||
if (blob_size == 0) {
|
||||
blob_size = n;
|
||||
}
|
||||
else if (n != blob_size) {
|
||||
LOGERR(1, "internal error: get_hashing_blob_nolock returned different blob size " << n << ", expected " << blob_size);
|
||||
}
|
||||
blobs.insert(blobs.end(), blob, blob + blob_size);
|
||||
}
|
||||
|
||||
|
|
|
@ -111,6 +111,37 @@ void StratumServer::on_block(const BlockTemplate& block)
|
|||
m_extraNonce.exchange(blobs_data->m_numClientsExpected);
|
||||
|
||||
blobs_data->m_blobSize = block.get_hashing_blobs(0, blobs_data->m_numClientsExpected, blobs_data->m_blobs, blobs_data->m_height, difficulty, sidechain_difficulty, blobs_data->m_seedHash, nonce_offset, blobs_data->m_templateId);
|
||||
|
||||
// Integrity checks
|
||||
if (blobs_data->m_blobSize < 76) {
|
||||
LOGERR(1, "internal error: get_hashing_blobs returned too small blobs (" << blobs_data->m_blobSize << " bytes)");
|
||||
}
|
||||
else if (blobs_data->m_blobs.size() != blobs_data->m_blobSize * num_connections) {
|
||||
LOGERR(1, "internal error: get_hashing_blobs returned wrong amount of data");
|
||||
}
|
||||
else if (num_connections > 1) {
|
||||
std::vector<uint64_t> blob_hashes;
|
||||
blob_hashes.reserve(num_connections);
|
||||
|
||||
const uint8_t* data = blobs_data->m_blobs.data();
|
||||
const size_t size = blobs_data->m_blobSize;
|
||||
|
||||
// Get first 8 bytes of the Merkle root hash from each blob
|
||||
for (size_t i = 0; i < num_connections; ++i) {
|
||||
blob_hashes.emplace_back(*reinterpret_cast<const uint64_t*>(data + i * size + 43));
|
||||
}
|
||||
|
||||
// Find duplicates
|
||||
std::sort(blob_hashes.begin(), blob_hashes.end());
|
||||
|
||||
for (uint32_t i = 1; i < num_connections; ++i) {
|
||||
if (blob_hashes[i - 1] == blob_hashes[i]) {
|
||||
LOGERR(1, "internal error: get_hashing_blobs returned two identical blobs");
|
||||
break;
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
blobs_data->m_target = std::max(difficulty.target(), sidechain_difficulty.target());
|
||||
|
||||
{
|
||||
|
|
Loading…
Reference in a new issue