Bitcoin Core Fuzz Coverage Report

Coverage Report

Created: 2026-03-24 13:57

next uncovered line (L), next uncovered region (R), next uncovered branch (B)
/root/bitcoin/src/blockencodings.cpp
Line
Count
Source
1
// Copyright (c) 2016-present The Bitcoin Core developers
2
// Distributed under the MIT software license, see the accompanying
3
// file COPYING or http://www.opensource.org/licenses/mit-license.php.
4
5
#include <blockencodings.h>
6
#include <chainparams.h>
7
#include <common/system.h>
8
#include <consensus/consensus.h>
9
#include <consensus/validation.h>
10
#include <crypto/sha256.h>
11
#include <crypto/siphash.h>
12
#include <logging.h>
13
#include <random.h>
14
#include <streams.h>
15
#include <txmempool.h>
16
#include <validation.h>
17
18
#include <unordered_map>
19
20
CBlockHeaderAndShortTxIDs::CBlockHeaderAndShortTxIDs(const CBlock& block, uint64_t nonce)
21
0
    : nonce(nonce),
22
0
      shorttxids(block.vtx.size() - 1),
23
0
      prefilledtxn(1),
24
0
      header(block)
25
0
{
26
0
    FillShortTxIDSelector();
27
    // TODO: Use our mempool prior to block acceptance to predictively fill more than just the coinbase
28
0
    prefilledtxn[0] = {0, block.vtx[0]};
29
0
    for (size_t i = 1; i < block.vtx.size(); i++) {
30
0
        const CTransaction& tx = *block.vtx[i];
31
0
        shorttxids[i - 1] = GetShortID(tx.GetWitnessHash());
32
0
    }
33
0
}
34
35
void CBlockHeaderAndShortTxIDs::FillShortTxIDSelector() const
36
0
{
37
0
    DataStream stream{};
38
0
    stream << header << nonce;
39
0
    CSHA256 hasher;
40
0
    hasher.Write((unsigned char*)&(*stream.begin()), stream.end() - stream.begin());
41
0
    uint256 shorttxidhash;
42
0
    hasher.Finalize(shorttxidhash.begin());
43
0
    m_hasher.emplace(shorttxidhash.GetUint64(0), shorttxidhash.GetUint64(1));
44
0
}
45
46
uint64_t CBlockHeaderAndShortTxIDs::GetShortID(const Wtxid& wtxid) const
47
0
{
48
0
    static_assert(SHORTTXIDS_LENGTH == 6, "shorttxids calculation assumes 6-byte shorttxids");
49
0
    return (*Assert(m_hasher))(wtxid.ToUint256()) & 0xffffffffffffL;
Line
Count
Source
113
0
#define Assert(val) inline_assertion_check<true>(val, std::source_location::current(), #val)
50
0
}
51
52
/* Reconstructing a compact block is in the hot-path for block relay,
53
 * so we want to do it as quickly as possible. Because this often
54
 * involves iterating over the entire mempool, we put all the data we
55
 * need (ie the wtxid and a reference to the actual transaction data)
56
 * in a vector and iterate over the vector directly. This allows optimal
57
 * CPU caching behaviour, at a cost of only 40 bytes per transaction.
58
 */
59
ReadStatus PartiallyDownloadedBlock::InitData(const CBlockHeaderAndShortTxIDs& cmpctblock, const std::vector<std::pair<Wtxid, CTransactionRef>>& extra_txn)
60
0
{
61
0
    LogDebug(BCLog::CMPCTBLOCK, "Initializing PartiallyDownloadedBlock for block %s using a cmpctblock of %u bytes\n", cmpctblock.header.GetHash().ToString(), GetSerializeSize(cmpctblock));
Line
Count
Source
115
0
#define LogDebug(category, ...) detail_LogIfCategoryAndLevelEnabled(category, BCLog::Level::Debug, __VA_ARGS__)
Line
Count
Source
106
0
    do {                                                               \
107
0
        if (util::log::ShouldLog((category), (level))) {               \
108
0
            bool rate_limit{level >= BCLog::Level::Info};              \
109
0
            Assume(!rate_limit); /*Only called with the levels below*/ \
Line
Count
Source
125
0
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
110
0
            LogPrintLevel_(category, level, rate_limit, __VA_ARGS__);  \
Line
Count
Source
89
0
#define LogPrintLevel_(category, level, should_ratelimit, ...) LogPrintFormatInternal(SourceLocation{__func__}, category, level, should_ratelimit, __VA_ARGS__)
111
0
        }                                                              \
112
0
    } while (0)
62
0
    if (cmpctblock.header.IsNull() || (cmpctblock.shorttxids.empty() && cmpctblock.prefilledtxn.empty()))
63
0
        return READ_STATUS_INVALID;
64
0
    if (cmpctblock.shorttxids.size() + cmpctblock.prefilledtxn.size() > MAX_BLOCK_WEIGHT / MIN_SERIALIZABLE_TRANSACTION_WEIGHT)
65
0
        return READ_STATUS_INVALID;
66
67
0
    if (!header.IsNull() || !txn_available.empty()) return READ_STATUS_INVALID;
68
69
0
    header = cmpctblock.header;
70
0
    txn_available.resize(cmpctblock.BlockTxCount());
71
72
0
    int32_t lastprefilledindex = -1;
73
0
    for (size_t i = 0; i < cmpctblock.prefilledtxn.size(); i++) {
74
0
        if (cmpctblock.prefilledtxn[i].tx->IsNull())
75
0
            return READ_STATUS_INVALID;
76
77
0
        lastprefilledindex += cmpctblock.prefilledtxn[i].index + 1; //index is a uint16_t, so can't overflow here
78
0
        if (lastprefilledindex > std::numeric_limits<uint16_t>::max())
79
0
            return READ_STATUS_INVALID;
80
0
        if ((uint32_t)lastprefilledindex > cmpctblock.shorttxids.size() + i) {
81
            // If we are inserting a tx at an index greater than our full list of shorttxids
82
            // plus the number of prefilled txn we've inserted, then we have txn for which we
83
            // have neither a prefilled txn or a shorttxid!
84
0
            return READ_STATUS_INVALID;
85
0
        }
86
0
        txn_available[lastprefilledindex] = cmpctblock.prefilledtxn[i].tx;
87
0
    }
88
0
    prefilled_count = cmpctblock.prefilledtxn.size();
89
90
    // Calculate map of txids -> positions and check mempool to see what we have (or don't)
91
    // Because well-formed cmpctblock messages will have a (relatively) uniform distribution
92
    // of short IDs, any highly-uneven distribution of elements can be safely treated as a
93
    // READ_STATUS_FAILED.
94
0
    std::unordered_map<uint64_t, uint16_t> shorttxids(cmpctblock.shorttxids.size());
95
0
    uint16_t index_offset = 0;
96
0
    for (size_t i = 0; i < cmpctblock.shorttxids.size(); i++) {
97
0
        while (txn_available[i + index_offset])
98
0
            index_offset++;
99
0
        shorttxids[cmpctblock.shorttxids[i]] = i + index_offset;
100
        // To determine the chance that the number of entries in a bucket exceeds N,
101
        // we use the fact that the number of elements in a single bucket is
102
        // binomially distributed (with n = the number of shorttxids S, and p =
103
        // 1 / the number of buckets), that in the worst case the number of buckets is
104
        // equal to S (due to std::unordered_map having a default load factor of 1.0),
105
        // and that the chance for any bucket to exceed N elements is at most
106
        // buckets * (the chance that any given bucket is above N elements).
107
        // Thus: P(max_elements_per_bucket > N) <= S * (1 - cdf(binomial(n=S,p=1/S), N)).
108
        // If we assume blocks of up to 16000, allowing 12 elements per bucket should
109
        // only fail once per ~1 million block transfers (per peer and connection).
110
0
        if (shorttxids.bucket_size(shorttxids.bucket(cmpctblock.shorttxids[i])) > 12)
111
0
            return READ_STATUS_FAILED;
112
0
    }
113
    // TODO: in the shortid-collision case, we should instead request both transactions
114
    // which collided. Falling back to full-block-request here is overkill.
115
0
    if (shorttxids.size() != cmpctblock.shorttxids.size())
116
0
        return READ_STATUS_FAILED; // Short ID collision
117
118
0
    std::vector<bool> have_txn(txn_available.size());
119
0
    {
120
0
    LOCK(pool->cs);
Line
Count
Source
266
0
#define LOCK(cs) UniqueLock UNIQUE_NAME(criticalblock)(MaybeCheckNotHeld(cs), #cs, __FILE__, __LINE__)
Line
Count
Source
11
0
#define UNIQUE_NAME(name) PASTE2(name, __COUNTER__)
Line
Count
Source
9
0
#define PASTE2(x, y) PASTE(x, y)
Line
Count
Source
8
0
#define PASTE(x, y) x ## y
121
0
    for (const auto& [wtxid, txit] : pool->txns_randomized) {
122
0
        uint64_t shortid = cmpctblock.GetShortID(wtxid);
123
0
        std::unordered_map<uint64_t, uint16_t>::iterator idit = shorttxids.find(shortid);
124
0
        if (idit != shorttxids.end()) {
125
0
            if (!have_txn[idit->second]) {
126
0
                txn_available[idit->second] = txit->GetSharedTx();
127
0
                have_txn[idit->second]  = true;
128
0
                mempool_count++;
129
0
            } else {
130
                // If we find two mempool txn that match the short id, just request it.
131
                // This should be rare enough that the extra bandwidth doesn't matter,
132
                // but eating a round-trip due to FillBlock failure would be annoying
133
0
                if (txn_available[idit->second]) {
134
0
                    txn_available[idit->second].reset();
135
0
                    mempool_count--;
136
0
                }
137
0
            }
138
0
        }
139
        // Though ideally we'd continue scanning for the two-txn-match-shortid case,
140
        // the performance win of an early exit here is too good to pass up and worth
141
        // the extra risk.
142
0
        if (mempool_count == shorttxids.size())
143
0
            break;
144
0
    }
145
0
    }
146
147
0
    for (size_t i = 0; i < extra_txn.size(); i++) {
148
0
        uint64_t shortid = cmpctblock.GetShortID(extra_txn[i].first);
149
0
        std::unordered_map<uint64_t, uint16_t>::iterator idit = shorttxids.find(shortid);
150
0
        if (idit != shorttxids.end()) {
151
0
            if (!have_txn[idit->second]) {
152
0
                txn_available[idit->second] = extra_txn[i].second;
153
0
                have_txn[idit->second]  = true;
154
0
                mempool_count++;
155
0
                extra_count++;
156
0
            } else {
157
                // If we find two mempool/extra txn that match the short id, just
158
                // request it.
159
                // This should be rare enough that the extra bandwidth doesn't matter,
160
                // but eating a round-trip due to FillBlock failure would be annoying
161
                // Note that we don't want duplication between extra_txn and mempool to
162
                // trigger this case, so we compare witness hashes first
163
0
                if (txn_available[idit->second] &&
164
0
                        txn_available[idit->second]->GetWitnessHash() != extra_txn[i].second->GetWitnessHash()) {
165
0
                    txn_available[idit->second].reset();
166
0
                    mempool_count--;
167
0
                    extra_count--;
168
0
                }
169
0
            }
170
0
        }
171
        // Though ideally we'd continue scanning for the two-txn-match-shortid case,
172
        // the performance win of an early exit here is too good to pass up and worth
173
        // the extra risk.
174
0
        if (mempool_count == shorttxids.size())
175
0
            break;
176
0
    }
177
178
0
    LogDebug(BCLog::CMPCTBLOCK, "Initialized PartiallyDownloadedBlock for block %s using a cmpctblock of %u bytes\n", cmpctblock.header.GetHash().ToString(), GetSerializeSize(cmpctblock));
Line
Count
Source
115
0
#define LogDebug(category, ...) detail_LogIfCategoryAndLevelEnabled(category, BCLog::Level::Debug, __VA_ARGS__)
Line
Count
Source
106
0
    do {                                                               \
107
0
        if (util::log::ShouldLog((category), (level))) {               \
108
0
            bool rate_limit{level >= BCLog::Level::Info};              \
109
0
            Assume(!rate_limit); /*Only called with the levels below*/ \
Line
Count
Source
125
0
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
110
0
            LogPrintLevel_(category, level, rate_limit, __VA_ARGS__);  \
Line
Count
Source
89
0
#define LogPrintLevel_(category, level, should_ratelimit, ...) LogPrintFormatInternal(SourceLocation{__func__}, category, level, should_ratelimit, __VA_ARGS__)
111
0
        }                                                              \
112
0
    } while (0)
179
180
0
    return READ_STATUS_OK;
181
0
}
182
183
bool PartiallyDownloadedBlock::IsTxAvailable(size_t index) const
184
0
{
185
0
    if (header.IsNull()) return false;
186
187
0
    assert(index < txn_available.size());
188
0
    return txn_available[index] != nullptr;
189
0
}
190
191
ReadStatus PartiallyDownloadedBlock::FillBlock(CBlock& block, const std::vector<CTransactionRef>& vtx_missing, bool segwit_active)
192
0
{
193
0
    if (header.IsNull()) return READ_STATUS_INVALID;
194
195
0
    block = header;
196
0
    block.vtx.resize(txn_available.size());
197
198
0
    size_t tx_missing_offset = 0;
199
0
    for (size_t i = 0; i < txn_available.size(); i++) {
200
0
        if (!txn_available[i]) {
201
0
            if (tx_missing_offset >= vtx_missing.size()) {
202
0
                return READ_STATUS_INVALID;
203
0
            }
204
0
            block.vtx[i] = vtx_missing[tx_missing_offset++];
205
0
        } else {
206
0
            block.vtx[i] = std::move(txn_available[i]);
207
0
        }
208
0
    }
209
210
    // Make sure we can't call FillBlock again.
211
0
    header.SetNull();
212
0
    txn_available.clear();
213
214
0
    if (vtx_missing.size() != tx_missing_offset) {
215
0
        return READ_STATUS_INVALID;
216
0
    }
217
218
    // Check for possible mutations early now that we have a seemingly good block
219
0
    IsBlockMutatedFn check_mutated{m_check_block_mutated_mock ? m_check_block_mutated_mock : IsBlockMutated};
220
0
    if (check_mutated(/*block=*/block, /*check_witness_root=*/segwit_active)) {
221
0
        return READ_STATUS_FAILED; // Possible Short ID collision
222
0
    }
223
224
0
    if (LogAcceptCategory(BCLog::CMPCTBLOCK, BCLog::Level::Debug)) {
225
0
        const uint256 hash{block.GetHash()};
226
0
        uint32_t tx_missing_size{0};
227
0
        for (const auto& tx : vtx_missing) tx_missing_size += tx->ComputeTotalSize();
228
0
        LogDebug(BCLog::CMPCTBLOCK, "Successfully reconstructed block %s with %u txn prefilled, %u txn from mempool (incl at least %u from extra pool) and %u txn (%u bytes) requested\n", hash.ToString(), prefilled_count, mempool_count, extra_count, vtx_missing.size(), tx_missing_size);
Line
Count
Source
115
0
#define LogDebug(category, ...) detail_LogIfCategoryAndLevelEnabled(category, BCLog::Level::Debug, __VA_ARGS__)
Line
Count
Source
106
0
    do {                                                               \
107
0
        if (util::log::ShouldLog((category), (level))) {               \
108
0
            bool rate_limit{level >= BCLog::Level::Info};              \
109
0
            Assume(!rate_limit); /*Only called with the levels below*/ \
Line
Count
Source
125
0
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
110
0
            LogPrintLevel_(category, level, rate_limit, __VA_ARGS__);  \
Line
Count
Source
89
0
#define LogPrintLevel_(category, level, should_ratelimit, ...) LogPrintFormatInternal(SourceLocation{__func__}, category, level, should_ratelimit, __VA_ARGS__)
111
0
        }                                                              \
112
0
    } while (0)
229
0
        if (vtx_missing.size() < 5) {
230
0
            for (const auto& tx : vtx_missing) {
231
0
                LogDebug(BCLog::CMPCTBLOCK, "Reconstructed block %s required tx %s\n", hash.ToString(), tx->GetHash().ToString());
Line
Count
Source
115
0
#define LogDebug(category, ...) detail_LogIfCategoryAndLevelEnabled(category, BCLog::Level::Debug, __VA_ARGS__)
Line
Count
Source
106
0
    do {                                                               \
107
0
        if (util::log::ShouldLog((category), (level))) {               \
108
0
            bool rate_limit{level >= BCLog::Level::Info};              \
109
0
            Assume(!rate_limit); /*Only called with the levels below*/ \
Line
Count
Source
125
0
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
110
0
            LogPrintLevel_(category, level, rate_limit, __VA_ARGS__);  \
Line
Count
Source
89
0
#define LogPrintLevel_(category, level, should_ratelimit, ...) LogPrintFormatInternal(SourceLocation{__func__}, category, level, should_ratelimit, __VA_ARGS__)
111
0
        }                                                              \
112
0
    } while (0)
232
0
            }
233
0
        }
234
0
    }
235
236
0
    return READ_STATUS_OK;
237
0
}