2019-05-25 18:06:42 +00:00
|
|
|
from ..merkle_minimal import merkleize_chunks, hash
|
2019-06-01 00:22:14 +00:00
|
|
|
from eth2spec.utils.ssz.ssz_typing import (
|
|
|
|
is_uint_type, is_bool_type, is_container_type,
|
|
|
|
is_list_kind, is_vector_kind,
|
|
|
|
read_vector_elem_type, read_elem_type,
|
|
|
|
uint_byte_size,
|
|
|
|
infer_input_type
|
|
|
|
)
|
2019-05-24 22:05:03 +00:00
|
|
|
|
|
|
|
# SSZ Serialization
|
2019-05-12 21:16:17 +00:00
|
|
|
# -----------------------------
|
|
|
|
|
2019-05-24 22:05:03 +00:00
|
|
|
BYTES_PER_LENGTH_OFFSET = 4
|
2019-05-12 21:16:17 +00:00
|
|
|
|
2019-05-27 16:01:46 +00:00
|
|
|
|
2019-05-25 18:06:42 +00:00
|
|
|
def is_basic_type(typ):
|
2019-05-27 20:18:34 +00:00
|
|
|
return is_uint_type(typ) or is_bool_type(typ)
|
2019-05-24 22:05:03 +00:00
|
|
|
|
2019-05-27 16:01:46 +00:00
|
|
|
|
2019-05-24 22:05:03 +00:00
|
|
|
def serialize_basic(value, typ):
|
2019-05-27 20:18:34 +00:00
|
|
|
if is_uint_type(typ):
|
2019-05-27 16:01:46 +00:00
|
|
|
return value.to_bytes(uint_byte_size(typ), 'little')
|
2019-05-30 13:42:42 +00:00
|
|
|
elif is_bool_type(typ):
|
2019-05-24 22:05:03 +00:00
|
|
|
if value:
|
|
|
|
return b'\x01'
|
|
|
|
else:
|
|
|
|
return b'\x00'
|
2019-05-30 13:42:42 +00:00
|
|
|
else:
|
|
|
|
raise Exception("Type not supported: {}".format(typ))
|
|
|
|
|
2019-06-01 00:22:14 +00:00
|
|
|
|
2019-05-30 13:42:42 +00:00
|
|
|
def deserialize_basic(value, typ):
|
|
|
|
if is_uint_type(typ):
|
|
|
|
return typ(int.from_bytes(value, 'little'))
|
|
|
|
elif is_bool_type(typ):
|
|
|
|
assert value in (b'\x00', b'\x01')
|
|
|
|
return True if value == b'\x01' else False
|
|
|
|
else:
|
|
|
|
raise Exception("Type not supported: {}".format(typ))
|
2019-05-24 22:05:03 +00:00
|
|
|
|
2019-05-27 16:01:46 +00:00
|
|
|
|
2019-05-25 18:06:42 +00:00
|
|
|
def is_fixed_size(typ):
|
|
|
|
if is_basic_type(typ):
|
|
|
|
return True
|
2019-05-27 20:18:34 +00:00
|
|
|
elif is_list_kind(typ):
|
2019-05-25 18:06:42 +00:00
|
|
|
return False
|
2019-05-27 20:18:34 +00:00
|
|
|
elif is_vector_kind(typ):
|
|
|
|
return is_fixed_size(read_vector_elem_type(typ))
|
|
|
|
elif is_container_type(typ):
|
2019-05-27 18:29:53 +00:00
|
|
|
return all(is_fixed_size(t) for t in typ.get_field_types())
|
2019-05-12 21:16:17 +00:00
|
|
|
else:
|
2019-05-25 18:06:42 +00:00
|
|
|
raise Exception("Type not supported: {}".format(typ))
|
|
|
|
|
2019-05-27 16:01:46 +00:00
|
|
|
|
|
|
|
@infer_input_type
|
2019-05-28 13:30:35 +00:00
|
|
|
def serialize(obj, typ=None):
|
2019-05-25 18:06:42 +00:00
|
|
|
if is_basic_type(typ):
|
|
|
|
return serialize_basic(obj, typ)
|
2019-05-27 20:18:34 +00:00
|
|
|
elif is_list_kind(typ) or is_vector_kind(typ):
|
|
|
|
return encode_series(obj, [read_elem_type(typ)]*len(obj))
|
|
|
|
elif is_container_type(typ):
|
2019-05-25 18:06:42 +00:00
|
|
|
return encode_series(obj.get_field_values(), typ.get_field_types())
|
|
|
|
else:
|
|
|
|
raise Exception("Type not supported: {}".format(typ))
|
2019-05-12 21:16:17 +00:00
|
|
|
|
2019-05-27 16:01:46 +00:00
|
|
|
|
2019-05-12 21:16:17 +00:00
|
|
|
def encode_series(values, types):
|
|
|
|
# bytes and bytesN are already in the right format.
|
|
|
|
if isinstance(values, bytes):
|
|
|
|
return values
|
|
|
|
|
|
|
|
# Recursively serialize
|
2019-05-28 18:33:12 +00:00
|
|
|
parts = [(is_fixed_size(types[i]), serialize(values[i], typ=types[i])) for i in range(len(values))]
|
2019-05-12 21:16:17 +00:00
|
|
|
|
|
|
|
# Compute and check lengths
|
|
|
|
fixed_lengths = [len(serialized) if constant_size else BYTES_PER_LENGTH_OFFSET
|
|
|
|
for (constant_size, serialized) in parts]
|
|
|
|
variable_lengths = [len(serialized) if not constant_size else 0
|
|
|
|
for (constant_size, serialized) in parts]
|
|
|
|
|
|
|
|
# Check if integer is not out of bounds (Python)
|
|
|
|
assert sum(fixed_lengths + variable_lengths) < 2 ** (BYTES_PER_LENGTH_OFFSET * 8)
|
|
|
|
|
|
|
|
# Interleave offsets of variable-size parts with fixed-size parts.
|
|
|
|
# Avoid quadratic complexity in calculation of offsets.
|
|
|
|
offset = sum(fixed_lengths)
|
|
|
|
variable_parts = []
|
|
|
|
fixed_parts = []
|
|
|
|
for (constant_size, serialized) in parts:
|
|
|
|
if constant_size:
|
|
|
|
fixed_parts.append(serialized)
|
|
|
|
else:
|
|
|
|
fixed_parts.append(offset.to_bytes(BYTES_PER_LENGTH_OFFSET, 'little'))
|
|
|
|
variable_parts.append(serialized)
|
|
|
|
offset += len(serialized)
|
|
|
|
|
|
|
|
# Return the concatenation of the fixed-size parts (offsets interleaved) with the variable-size parts
|
|
|
|
return b''.join(fixed_parts + variable_parts)
|
2019-05-24 22:05:03 +00:00
|
|
|
|
2019-05-27 16:01:46 +00:00
|
|
|
|
2019-05-25 18:06:42 +00:00
|
|
|
# SSZ Hash-tree-root
|
|
|
|
# -----------------------------
|
|
|
|
|
2019-05-27 16:01:46 +00:00
|
|
|
|
2019-05-25 18:06:42 +00:00
|
|
|
def pack(values, subtype):
|
|
|
|
if isinstance(values, bytes):
|
|
|
|
return values
|
|
|
|
return b''.join([serialize_basic(value, subtype) for value in values])
|
|
|
|
|
2019-05-27 16:01:46 +00:00
|
|
|
|
2019-05-25 18:06:42 +00:00
|
|
|
def chunkify(bytez):
|
|
|
|
bytez += b'\x00' * (-len(bytez) % 32)
|
|
|
|
return [bytez[i:i + 32] for i in range(0, len(bytez), 32)]
|
|
|
|
|
2019-05-27 16:01:46 +00:00
|
|
|
|
2019-05-25 18:06:42 +00:00
|
|
|
def mix_in_length(root, length):
|
|
|
|
return hash(root + length.to_bytes(32, 'little'))
|
|
|
|
|
2019-05-27 16:01:46 +00:00
|
|
|
|
2019-05-28 18:33:12 +00:00
|
|
|
def is_bottom_layer_kind(typ):
|
|
|
|
return (
|
|
|
|
is_basic_type(typ) or
|
|
|
|
(is_list_kind(typ) or is_vector_kind(typ)) and is_basic_type(read_elem_type(typ))
|
|
|
|
)
|
|
|
|
|
|
|
|
|
2019-05-27 16:01:46 +00:00
|
|
|
@infer_input_type
|
2019-05-28 18:33:12 +00:00
|
|
|
def get_typed_values(obj, typ=None):
|
|
|
|
if is_container_type(typ):
|
|
|
|
return obj.get_typed_values()
|
2019-05-27 20:18:34 +00:00
|
|
|
elif is_list_kind(typ) or is_vector_kind(typ):
|
2019-05-28 18:33:12 +00:00
|
|
|
elem_type = read_elem_type(typ)
|
2019-05-30 13:42:42 +00:00
|
|
|
return list(zip(obj, [elem_type] * len(obj)))
|
2019-05-28 18:33:12 +00:00
|
|
|
else:
|
|
|
|
raise Exception("Invalid type")
|
|
|
|
|
|
|
|
|
|
|
|
@infer_input_type
|
|
|
|
def hash_tree_root(obj, typ=None):
|
|
|
|
if is_bottom_layer_kind(typ):
|
|
|
|
data = serialize_basic(obj, typ) if is_basic_type(typ) else pack(obj, read_elem_type(typ))
|
|
|
|
leaves = chunkify(data)
|
|
|
|
else:
|
|
|
|
fields = get_typed_values(obj, typ=typ)
|
|
|
|
leaves = [hash_tree_root(field_value, typ=field_typ) for field_value, field_typ in fields]
|
|
|
|
if is_list_kind(typ):
|
|
|
|
return mix_in_length(merkleize_chunks(leaves), len(obj))
|
2019-05-25 18:06:42 +00:00
|
|
|
else:
|
2019-05-28 18:33:12 +00:00
|
|
|
return merkleize_chunks(leaves)
|
2019-05-25 18:06:42 +00:00
|
|
|
|
2019-05-27 16:01:46 +00:00
|
|
|
|
|
|
|
@infer_input_type
|
|
|
|
def signing_root(obj, typ):
|
2019-05-27 20:18:34 +00:00
|
|
|
assert is_container_type(typ)
|
2019-05-27 21:40:05 +00:00
|
|
|
# ignore last field
|
2019-05-28 18:33:12 +00:00
|
|
|
leaves = [hash_tree_root(field_value, typ=field_typ) for field_value, field_typ in obj.get_typed_values()[:-1]]
|
2019-05-25 18:06:42 +00:00
|
|
|
return merkleize_chunks(chunkify(b''.join(leaves)))
|