Adds missing functions

This commit is contained in:
Carl Beekhuizen 2019-05-20 10:50:07 +02:00
parent 725dcf27f5
commit cfed4bb41c
No known key found for this signature in database
GPG Key ID: D05CA176D0020646
4 changed files with 50 additions and 85 deletions

View File

@ -5,7 +5,7 @@ from argparse import ArgumentParser
from typing import Tuple, List
IMPORTS = '''from typing import (
PHASE0_IMPORTS = '''from typing import (
Any,
Dict,
List,
@ -27,6 +27,31 @@ from eth2spec.utils.bls_stub import (
from eth2spec.utils.hash_function import hash
'''
PHASE1_IMPORTS = '''from typing import (
Any,
Dict,
List,
NewType,
Tuple,
)
from eth2spec.utils.minimal_ssz import (
SSZType,
hash_tree_root,
signing_root,
type_of,
empty,
)
from eth2spec.utils.bls_stub import (
bls_aggregate_pubkeys,
bls_verify,
bls_verify_multiple,
)
from eth2spec.utils.hash_function import hash
import math
'''
NEW_TYPE_DEFINITIONS = '''
Slot = NewType('Slot', int) # uint64
Epoch = NewType('Epoch', int) # uint64
@ -82,74 +107,8 @@ def apply_constants_preset(preset: Dict[str, Any]):
init_SSZ_types()
'''
def split_and_label(regex_pattern: str, text: str) -> List[str]:
'''
Splits a string based on regex, but down not remove the matched text.
It subsequently labels the matches with their match
'''
find_pattern = r'''%s.*?(?=%s|$)''' % (regex_pattern, regex_pattern)
matches = re.findall(find_pattern, text, re.DOTALL)
return list(map(lambda x: [re.match(regex_pattern, x).group(0), x], matches))
def inserter(oldfile: str, newfile: str) -> Tuple[str, str]:
'''
Searches for occurrences of @LabelName in oldfile and replaces them with instances of code wraped as follows:
# begin insert @LabelName
def foo(bar):
return bar
# end insert @LabelName
'''
new_insert_objects = re.split(r"(# begin insert |# end insert @[\w\d_]*\n)", newfile)
# Retrieve label from insert objects
def get_labeled_object(labeled_text):
label = re.match(r"@[\w\d_]*\n", labeled_text)
if label is not None:
label = label.group(0)
labeled_text = re.sub(label, '', labeled_text)
return {'label': label, 'text': labeled_text}
new_insert_objects = map(get_labeled_object, new_insert_objects)
# Find and replace labels
newfile = ""
for item in new_insert_objects:
if item['label'] is not None:
oldfile, insertions = re.subn('# %s' % item['label'], item['text'], oldfile)
if insertions == 0:
newfile.join('# begin insert %s/n%s# end insert %s' % (item['label'], item['text'], item['label']))
elif re.match(r"(# begin insert |# end insert )", item['text']) is None:
newfile += item['text']
return oldfile, newfile
def merger(oldfile:str, newfile:str) -> str:
'''
Seeks out functions and objects in new and old files.
Replaces old objects with new ones if they exist.
'''
object_regex = r'''(?:\n[@\w]+[\s\w]*[='" "\.\w]*)|(?:\s{4}global_vars\["\w+"\])'''
ssz_object_regex = r'''(?:\w+|\s{4}global_vars\["\w+"\]) = SSZType\(\{\n'''
old_objects = split_and_label(object_regex, oldfile)
new_objects = split_and_label(object_regex, newfile)
for new_item in new_objects:
found_old = False
for index, old_item in enumerate(old_objects):
if old_item[0] == new_item[0]:
ssz_object_match = re.match(ssz_object_regex, new_item[1])
if ssz_object_match is not None:
new_item[1] = re.sub(ssz_object_regex, '', new_item[1])
old_item[1] = re.sub(r'\n\w*\}\)', '', old_item[1])
old_item[1] += new_item[1]
else:
old_item[1] = new_item[1]
found_old = True
old_objects[index] = old_item
break
if not found_old:
old_objects.append(new_item)
return ''.join(elem for elem in map(lambda x: x[1], old_objects))
def objects_to_spec(functions, constants, ssz_objects):
def objects_to_spec(functions, constants, ssz_objects, imports):
functions_spec = '\n\n'.join(functions.values())
constants_spec = '\n'.join(map(lambda x: '%s = %s' % (x, constants[x]),constants))
ssz_objects_instantiation_spec = '\n'.join(map(lambda x: '%s = SSZType(%s)' % (x, ssz_objects[x][:-1]), ssz_objects))
@ -160,7 +119,7 @@ def objects_to_spec(functions, constants, ssz_objects):
+ ssz_objects_reinitialization_spec
)
return (
IMPORTS
imports
+ '\n' + NEW_TYPE_DEFINITIONS
+ '\n' + constants_spec
+ '\n' + ssz_objects_instantiation_spec
@ -194,7 +153,7 @@ def combine_ssz_objects(old_objects, new_objects):
def build_phase0_spec(sourcefile, outfile=None):
functions, constants, ssz_objects = function_puller.get_spec(sourcefile)
spec = objects_to_spec(functions, constants, ssz_objects)
spec = objects_to_spec(functions, constants, ssz_objects, PHASE0_IMPORTS)
if outfile is not None:
with open(outfile, 'w') as out:
out.write(spec)
@ -208,7 +167,7 @@ def build_phase1_spec(phase0_sourcefile, phase1_sourcefile, outfile=None):
functions = combine_functions(phase0_functions, phase1_functions)
constants = combine_constants(phase0_constants, phase1_constants)
ssz_objects = combine_functions(phase0_ssz_objects, phase1_ssz_objects)
spec = objects_to_spec(functions, constants, ssz_objects)
spec = objects_to_spec(functions, constants, ssz_objects, PHASE1_IMPORTS)
if outfile is not None:
with open(outfile, 'w') as out:
out.write(spec)

View File

@ -15,7 +15,6 @@ def get_spec(file_name: str):
constants = {}
ssz_objects = defaultdict(str)
function_matcher = re.compile(FUNCTION_REGEX)
# type_defs = []
for linenum, line in enumerate(open(file_name).readlines()):
line = line.rstrip()
if pulling_from is None and len(line) > 0 and line[0] == '#' and line[-1] == '`':
@ -26,9 +25,6 @@ def get_spec(file_name: str):
elif line[:3] == '```':
pulling_from = None
else:
# # Handle SSZObjects
# if pulling_from == linenum and line == '{':
# code_lines.append('%s = SSZType({' % current_name)
# Handle function definitions
if pulling_from is not None:
match = function_matcher.match(line)

View File

@ -31,7 +31,7 @@
- [`BeaconState`](#beaconstate)
- [`BeaconBlockBody`](#beaconblockbody)
- [Helpers](#helpers)
- [`typeof`](#typeof)
- [`type_of`](#type_of)
- [`empty`](#empty)
- [`get_crosslink_chunk_count`](#get_crosslink_chunk_count)
- [`get_custody_chunk_bit`](#get_custody_chunk_bit)
@ -259,13 +259,13 @@ Add the following fields to the end of the specified container objects. Fields w
## Helpers
### `typeof`
### `type_of`
The `typeof` function accepts and SSZ object as a single input and returns the corresponding SSZ type.
The `type_of` function accepts an SSZ object as a single input and returns the corresponding SSZ type.
### `empty`
The `empty` function accepts and SSZ type as input and returns an object of that type with all fields initialized to default values.
The `empty` function accepts an SSZ type as input and returns an object of that type with all fields initialized to default values.
### `get_crosslink_chunk_count`
@ -327,7 +327,7 @@ def get_validators_custody_reveal_period(state: BeaconState,
```python
def replace_empty_or_append(list: List[Any], new_element: Any) -> int:
for i in range(len(list)):
if list[i] == empty(typeof(new_element)):
if list[i] == empty(type_of(new_element)):
list[i] = new_element
return i
list.append(new_element)
@ -492,7 +492,7 @@ def process_chunk_challenge(state: BeaconState,
record.chunk_index != challenge.chunk_index
)
# Verify depth
depth = math.log2(next_power_of_two(get_custody_chunk_count(challenge.attestation)))
depth = math.ceil(math.log2(get_custody_chunk_count(challenge.attestation)))
assert challenge.chunk_index < 2**depth
# Add new chunk challenge record
new_record = CustodyChunkChallengeRecord(
@ -580,7 +580,7 @@ def process_bit_challenge(state: BeaconState,
deadline=get_current_epoch(state) + CUSTODY_RESPONSE_DEADLINE,
data_root=challenge.attestation.data.crosslink.data_root,
chunk_count=chunk_count,
chunk_bits_merkle_root=merkle_root(pad_to_power_of_2((challenge.chunk_bits))),
chunk_bits_merkle_root=hash_tree_root(challenge.chunk_bits),
responder_key=challenge.responder_key,
)
replace_empty_or_append(state.custody_bit_challenge_records, new_record)
@ -646,7 +646,7 @@ def process_bit_challenge_response(state: BeaconState,
assert verify_merkle_branch(
leaf=hash_tree_root(response.chunk),
branch=response.data_branch,
depth=math.log2(next_power_of_two(challenge.chunk_count)),
depth=math.ceil(math.log2(challenge.chunk_count)),
index=response.chunk_index,
root=challenge.data_root,
)
@ -654,13 +654,13 @@ def process_bit_challenge_response(state: BeaconState,
assert verify_merkle_branch(
leaf=response.chunk_bits_leaf,
branch=response.chunk_bits_branch,
depth=math.log2(next_power_of_two(challenge.chunk_count) // 256),
depth=math.ceil(math.log2(challenge.chunk_count)) >> 8,
index=response.chunk_index // 256,
root=challenge.chunk_bits_merkle_root
)
# Verify the chunk bit does not match the challenge chunk bit
assert (get_custody_chunk_bit(challenge.responder_key, response.chunk)
!= get_bitfield_bit(challenge.chunk_bits_leaf, response.chunk_index % 256))
!= get_bitfield_bit(challenge.chunk_bits_leaf, response.chunk_index % 256))
# Clear the challenge
records = state.custody_bit_challenge_records
records[records.index(challenge)] = CustodyBitChallengeRecord()

View File

@ -58,6 +58,16 @@ class Vector():
return self.length
def type_of(obj):
return obj.__class__
def empty(obj):
for field in obj.fields:
field = get_zero_value(field)
return obj
def is_basic(typ):
# if not a string, it is a complex, and cannot be basic
if not isinstance(typ, str):