# beacon_chain # Copyright (c) 2018-2024 Status Research & Development GmbH # Licensed and distributed under either of # * MIT license (license terms in the root directory or at https://opensource.org/licenses/MIT). # * Apache v2 license (license terms in the root directory or at https://www.apache.org/licenses/LICENSE-2.0). # at your option. This file may not be copied, modified, or distributed except according to those terms. {.push raises: [].} import std/tables, eth/p2p/discoveryv5/[node], ../spec/[helpers, eip7594_helpers] from std/sequtils import mapIt from std/strutils import join const MaxDataColumns = 3 * SLOTS_PER_EPOCH * NUMBER_OF_COLUMNS ## Same limit as `MaxOrphans` in `block_quarantine` ## data columns may arrive before an orphan is tagged `columnless` type DataColumnQuarantine* = object data_columns*: OrderedTable[(Eth2Digest, ColumnIndex), ref DataColumnSidecar] supernode*: bool nodeid*: NodeId onDataColumnSidecarCallback*: OnDataColumnSidecarCallback DataColumnFetchRecord* = object block_root*: Eth2Digest indices*: seq[ColumnIndex] OnDataColumnSidecarCallback = proc(data: DataColumnSidecar) {.gcsafe, raises: [].} func shortLog*(x: seq[ColumnIndex]): string = "<" & x.mapIt($it).join(", ") & ">" func shortLog*(x: seq[DataColumnFetchRecord]): string = "[" & x.mapIt(shortLog(it.block_root) & shortLog(it.indices)).join(", ") & "]" func put*(quarantine: var DataColumnQuarantine, dataColumnSidecar: ref DataColumnSidecar) = if quarantine.data_columns.lenu64 >= MaxDataColumns: # FIFO if full. For example, sync manager and request manager can race to # put data columns in at the same time, so one gets blob insert -> block resolve # -> data columns insert sequence, which leaves garbage data columns. # # This also therefore automatically garbage-collects otherwise valid garbage # data columns which are correctly signed, point to either correct block roots or a # block root which isn't ever seen, and then are for any reason simply never # used. var oldest_column_key: (Eth2Digest, ColumnIndex) for k in quarantine.data_columns.keys: oldest_column_key = k break quarantine.data_columns.del(oldest_column_key) let block_root = hash_tree_root(dataColumnSidecar.signed_block_header.message) discard quarantine.data_columns.hasKeyOrPut( (block_root, dataColumnSidecar.index), dataColumnSidecar) func hasDataColumn*( quarantine: DataColumnQuarantine, slot: Slot, proposer_index: uint64, index: ColumnIndex): bool = for data_column_sidecar in quarantine.data_columns.values: template block_header: untyped = data_column_sidecar.signed_block_header.message if block_header.slot == slot and block_header.proposer_index == proposer_index and data_column_sidecar.index == index: return true false func accumulateDataColumns*(quarantine: DataColumnQuarantine, blck: deneb.SignedBeaconBlock | electra.SignedBeaconBlock): seq[ColumnIndex] = # This method copies the DataColumns that were received via # gossip and are accumulated to an array var indices: seq[ColumnIndex] for i in 0..= (NUMBER_OF_COLUMNS div 2): return true else: for i in localCustodyColumns: if (blck.root, ColumnIndex i) notin quarantine.data_columns: return false true func dataColumnFetchRecord*(quarantine: DataColumnQuarantine, blck: deneb.SignedBeaconBlock | electra.SignedBeaconBlock): DataColumnFetchRecord = let localSubnetCount = if quarantine.supernode: DATA_COLUMN_SIDECAR_SUBNET_COUNT.uint64 else: CUSTODY_REQUIREMENT.uint64 localCustodyColumns = get_custody_columns(quarantine.nodeid, max(SAMPLES_PER_SLOT.uint64, localSubnetCount)) var indices: seq[ColumnIndex] for i in localCustodyColumns: let idx = ColumnIndex(i) if not quarantine.data_columns.hasKey( (blck.root, idx)): indices.add(idx) DataColumnFetchRecord(block_root: blck.root, indices: indices) func init*( T: type DataColumnQuarantine): T = T()