feat_: introduce forward error correction in segmentation layer

closes: #4330
This commit is contained in:
Patryk Osmaczko 2024-04-04 19:46:51 +02:00 committed by osmaczko
parent 4f2adc0ced
commit 9e5462eb9e
8 changed files with 667 additions and 165 deletions

View File

@ -0,0 +1,365 @@
package common
import (
"bytes"
"math"
"time"
"github.com/golang/protobuf/proto"
"github.com/jinzhu/copier"
"github.com/klauspost/reedsolomon"
"github.com/pkg/errors"
"go.uber.org/zap"
"github.com/status-im/status-go/eth-node/crypto"
"github.com/status-im/status-go/eth-node/types"
"github.com/status-im/status-go/protocol/protobuf"
v1protocol "github.com/status-im/status-go/protocol/v1"
)
var ErrMessageSegmentsIncomplete = errors.New("message segments incomplete")
var ErrMessageSegmentsAlreadyCompleted = errors.New("message segments already completed")
var ErrMessageSegmentsInvalidCount = errors.New("invalid segments count")
var ErrMessageSegmentsHashMismatch = errors.New("hash of entire payload does not match")
var ErrMessageSegmentsInvalidParity = errors.New("invalid parity segments")
const (
segmentsParityRate = 0.125
segmentsReedsolomonMaxCount = 256
)
type SegmentMessage struct {
*protobuf.SegmentMessage
}
func (s *SegmentMessage) IsValid() bool {
return s.SegmentsCount >= 2 || s.ParitySegmentsCount > 0
}
func (s *SegmentMessage) IsParityMessage() bool {
return s.SegmentsCount == 0 && s.ParitySegmentsCount > 0
}
func (s *MessageSender) segmentMessage(newMessage *types.NewMessage) ([]*types.NewMessage, error) {
// We set the max message size to 3/4 of the allowed message size, to leave
// room for segment message metadata.
newMessages, err := segmentMessage(newMessage, int(s.transport.MaxMessageSize()/4*3))
s.logger.Debug("message segmented", zap.Int("segments", len(newMessages)))
return newMessages, err
}
func replicateMessageWithNewPayload(message *types.NewMessage, payload []byte) (*types.NewMessage, error) {
copy := &types.NewMessage{}
err := copier.Copy(copy, message)
if err != nil {
return nil, err
}
copy.Payload = payload
copy.PowTarget = calculatePoW(payload)
return copy, nil
}
// Segments message into smaller chunks if the size exceeds segmentSize.
func segmentMessage(newMessage *types.NewMessage, segmentSize int) ([]*types.NewMessage, error) {
if len(newMessage.Payload) <= segmentSize {
return []*types.NewMessage{newMessage}, nil
}
entireMessageHash := crypto.Keccak256(newMessage.Payload)
entirePayloadSize := len(newMessage.Payload)
segmentsCount := int(math.Ceil(float64(entirePayloadSize) / float64(segmentSize)))
paritySegmentsCount := int(math.Floor(float64(segmentsCount) * segmentsParityRate))
segmentPayloads := make([][]byte, segmentsCount+paritySegmentsCount)
segmentMessages := make([]*types.NewMessage, segmentsCount)
for start, index := 0, 0; start < entirePayloadSize; start += segmentSize {
end := start + segmentSize
if end > entirePayloadSize {
end = entirePayloadSize
}
segmentPayload := newMessage.Payload[start:end]
segmentWithMetadata := &protobuf.SegmentMessage{
EntireMessageHash: entireMessageHash,
Index: uint32(index),
SegmentsCount: uint32(segmentsCount),
Payload: segmentPayload,
}
marshaledSegmentWithMetadata, err := proto.Marshal(segmentWithMetadata)
if err != nil {
return nil, err
}
segmentMessage, err := replicateMessageWithNewPayload(newMessage, marshaledSegmentWithMetadata)
if err != nil {
return nil, err
}
segmentPayloads[index] = segmentPayload
segmentMessages[index] = segmentMessage
index++
}
// Skip reedsolomon if the combined total of data and parity segments exceeds the predefined limit of segmentsReedsolomonMaxCount.
// Exceeding this limit necessitates shard sizes to be multiples of 64, which are incompatible with clients that do not support forward error correction.
if paritySegmentsCount == 0 || segmentsCount+paritySegmentsCount > segmentsReedsolomonMaxCount {
return segmentMessages, nil
}
enc, err := reedsolomon.New(segmentsCount, paritySegmentsCount)
if err != nil {
return nil, err
}
// Align the size of the last segment payload.
lastSegmentPayload := segmentPayloads[segmentsCount-1]
segmentPayloads[segmentsCount-1] = make([]byte, segmentSize)
copy(segmentPayloads[segmentsCount-1], lastSegmentPayload)
// Make space for parity data.
for i := segmentsCount; i < segmentsCount+paritySegmentsCount; i++ {
segmentPayloads[i] = make([]byte, segmentSize)
}
err = enc.Encode(segmentPayloads)
if err != nil {
return nil, err
}
// Create parity messages.
for i, index := segmentsCount, 0; i < segmentsCount+paritySegmentsCount; i++ {
segmentWithMetadata := &protobuf.SegmentMessage{
EntireMessageHash: entireMessageHash,
SegmentsCount: 0, // indicates parity message
ParitySegmentIndex: uint32(index),
ParitySegmentsCount: uint32(paritySegmentsCount),
Payload: segmentPayloads[i],
}
marshaledSegmentWithMetadata, err := proto.Marshal(segmentWithMetadata)
if err != nil {
return nil, err
}
segmentMessage, err := replicateMessageWithNewPayload(newMessage, marshaledSegmentWithMetadata)
if err != nil {
return nil, err
}
segmentMessages = append(segmentMessages, segmentMessage)
index++
}
return segmentMessages, nil
}
// SegmentationLayerV1 reconstructs the message only when all segments have been successfully retrieved.
// It lacks the capability to perform forward error correction.
// Kept to test forward compatibility.
func (s *MessageSender) handleSegmentationLayerV1(message *v1protocol.StatusMessage) error {
logger := s.logger.With(zap.String("site", "handleSegmentationLayerV1")).With(zap.String("hash", types.HexBytes(message.TransportLayer.Hash).String()))
segmentMessage := &SegmentMessage{
SegmentMessage: &protobuf.SegmentMessage{},
}
err := proto.Unmarshal(message.TransportLayer.Payload, segmentMessage.SegmentMessage)
if err != nil {
return errors.Wrap(err, "failed to unmarshal SegmentMessage")
}
logger.Debug("handling message segment", zap.String("EntireMessageHash", types.HexBytes(segmentMessage.EntireMessageHash).String()),
zap.Uint32("Index", segmentMessage.Index), zap.Uint32("SegmentsCount", segmentMessage.SegmentsCount))
alreadyCompleted, err := s.persistence.IsMessageAlreadyCompleted(segmentMessage.EntireMessageHash)
if err != nil {
return err
}
if alreadyCompleted {
return ErrMessageSegmentsAlreadyCompleted
}
if segmentMessage.SegmentsCount < 2 {
return ErrMessageSegmentsInvalidCount
}
err = s.persistence.SaveMessageSegment(segmentMessage, message.TransportLayer.SigPubKey, time.Now().Unix())
if err != nil {
return err
}
segments, err := s.persistence.GetMessageSegments(segmentMessage.EntireMessageHash, message.TransportLayer.SigPubKey)
if err != nil {
return err
}
if len(segments) != int(segmentMessage.SegmentsCount) {
return ErrMessageSegmentsIncomplete
}
// Combine payload
var entirePayload bytes.Buffer
for _, segment := range segments {
_, err := entirePayload.Write(segment.Payload)
if err != nil {
return errors.Wrap(err, "failed to write segment payload")
}
}
// Sanity check
entirePayloadHash := crypto.Keccak256(entirePayload.Bytes())
if !bytes.Equal(entirePayloadHash, segmentMessage.EntireMessageHash) {
return ErrMessageSegmentsHashMismatch
}
err = s.persistence.CompleteMessageSegments(segmentMessage.EntireMessageHash, message.TransportLayer.SigPubKey, time.Now().Unix())
if err != nil {
return err
}
message.TransportLayer.Payload = entirePayload.Bytes()
return nil
}
// SegmentationLayerV2 is capable of reconstructing the message from both complete and partial sets of data segments.
// It has capability to perform forward error correction.
func (s *MessageSender) handleSegmentationLayerV2(message *v1protocol.StatusMessage) error {
logger := s.logger.With(zap.String("site", "handleSegmentationLayerV2")).With(zap.String("hash", types.HexBytes(message.TransportLayer.Hash).String()))
segmentMessage := &SegmentMessage{
SegmentMessage: &protobuf.SegmentMessage{},
}
err := proto.Unmarshal(message.TransportLayer.Payload, segmentMessage.SegmentMessage)
if err != nil {
return errors.Wrap(err, "failed to unmarshal SegmentMessage")
}
logger.Debug("handling message segment",
zap.String("EntireMessageHash", types.HexBytes(segmentMessage.EntireMessageHash).String()),
zap.Uint32("Index", segmentMessage.Index),
zap.Uint32("SegmentsCount", segmentMessage.SegmentsCount),
zap.Uint32("ParitySegmentIndex", segmentMessage.ParitySegmentIndex),
zap.Uint32("ParitySegmentsCount", segmentMessage.ParitySegmentsCount))
alreadyCompleted, err := s.persistence.IsMessageAlreadyCompleted(segmentMessage.EntireMessageHash)
if err != nil {
return err
}
if alreadyCompleted {
return ErrMessageSegmentsAlreadyCompleted
}
if !segmentMessage.IsValid() {
return ErrMessageSegmentsInvalidCount
}
err = s.persistence.SaveMessageSegment(segmentMessage, message.TransportLayer.SigPubKey, time.Now().Unix())
if err != nil {
return err
}
segments, err := s.persistence.GetMessageSegments(segmentMessage.EntireMessageHash, message.TransportLayer.SigPubKey)
if err != nil {
return err
}
if len(segments) == 0 {
return errors.New("unexpected state: no segments found after save operation") // This should theoretically never occur.
}
firstSegmentMessage := segments[0]
lastSegmentMessage := segments[len(segments)-1]
// First segment message must not be a parity message.
if firstSegmentMessage.IsParityMessage() || len(segments) != int(firstSegmentMessage.SegmentsCount) {
return ErrMessageSegmentsIncomplete
}
payloads := make([][]byte, firstSegmentMessage.SegmentsCount+lastSegmentMessage.ParitySegmentsCount)
payloadSize := len(firstSegmentMessage.Payload)
restoreUsingParityData := lastSegmentMessage.IsParityMessage()
if !restoreUsingParityData {
for i, segment := range segments {
payloads[i] = segment.Payload
}
} else {
enc, err := reedsolomon.New(int(firstSegmentMessage.SegmentsCount), int(lastSegmentMessage.ParitySegmentsCount))
if err != nil {
return err
}
var lastNonParitySegmentPayload []byte
for _, segment := range segments {
if !segment.IsParityMessage() {
if segment.Index == firstSegmentMessage.SegmentsCount-1 {
// Ensure last segment is aligned to payload size, as it is required by reedsolomon.
payloads[segment.Index] = make([]byte, payloadSize)
copy(payloads[segment.Index], segment.Payload)
lastNonParitySegmentPayload = segment.Payload
} else {
payloads[segment.Index] = segment.Payload
}
} else {
payloads[firstSegmentMessage.SegmentsCount+segment.ParitySegmentIndex] = segment.Payload
}
}
err = enc.Reconstruct(payloads)
if err != nil {
return err
}
ok, err := enc.Verify(payloads)
if err != nil {
return err
}
if !ok {
return ErrMessageSegmentsInvalidParity
}
if lastNonParitySegmentPayload != nil {
payloads[firstSegmentMessage.SegmentsCount-1] = lastNonParitySegmentPayload // Bring back last segment with original length.
}
}
// Combine payload.
var entirePayload bytes.Buffer
for i := 0; i < int(firstSegmentMessage.SegmentsCount); i++ {
_, err := entirePayload.Write(payloads[i])
if err != nil {
return errors.Wrap(err, "failed to write segment payload")
}
}
// Sanity check.
entirePayloadHash := crypto.Keccak256(entirePayload.Bytes())
if !bytes.Equal(entirePayloadHash, segmentMessage.EntireMessageHash) {
return ErrMessageSegmentsHashMismatch
}
err = s.persistence.CompleteMessageSegments(segmentMessage.EntireMessageHash, message.TransportLayer.SigPubKey, time.Now().Unix())
if err != nil {
return err
}
message.TransportLayer.Payload = entirePayload.Bytes()
return nil
}
func (s *MessageSender) CleanupSegments() error {
monthAgo := time.Now().AddDate(0, -1, 0).Unix()
err := s.persistence.RemoveMessageSegmentsOlderThan(monthAgo)
if err != nil {
return err
}
err = s.persistence.RemoveMessageSegmentsCompletedOlderThan(monthAgo)
if err != nil {
return err
}
return nil
}

View File

@ -0,0 +1,205 @@
package common
import (
"fmt"
"math"
"testing"
"github.com/stretchr/testify/suite"
"go.uber.org/zap"
"golang.org/x/exp/slices"
"github.com/status-im/status-go/appdatabase"
"github.com/status-im/status-go/eth-node/crypto"
"github.com/status-im/status-go/eth-node/types"
"github.com/status-im/status-go/protocol/sqlite"
"github.com/status-im/status-go/protocol/v1"
"github.com/status-im/status-go/t/helpers"
)
func TestMessageSegmentationSuite(t *testing.T) {
suite.Run(t, new(MessageSegmentationSuite))
}
type MessageSegmentationSuite struct {
suite.Suite
sender *MessageSender
testPayload []byte
logger *zap.Logger
}
func (s *MessageSegmentationSuite) SetupSuite() {
s.testPayload = make([]byte, 1000)
for i := 0; i < 1000; i++ {
s.testPayload[i] = byte(i)
}
}
func (s *MessageSegmentationSuite) SetupTest() {
identity, err := crypto.GenerateKey()
s.Require().NoError(err)
database, err := helpers.SetupTestMemorySQLDB(appdatabase.DbInitializer{})
s.Require().NoError(err)
err = sqlite.Migrate(database)
s.Require().NoError(err)
s.logger, err = zap.NewDevelopment()
s.Require().NoError(err)
s.sender, err = NewMessageSender(
identity,
database,
nil,
nil,
s.logger,
FeatureFlags{},
)
s.Require().NoError(err)
}
func (s *MessageSegmentationSuite) SetupSubTest() {
s.SetupTest()
}
func (s *MessageSegmentationSuite) TestHandleSegmentationLayer() {
testCases := []struct {
name string
segmentsCount int
expectedParitySegmentsCount int
retrievedSegments []int
retrievedParitySegments []int
segmentationLayerV1ShouldSucceed bool
segmentationLayerV2ShouldSucceed bool
}{
{
name: "all segments retrieved",
segmentsCount: 2,
expectedParitySegmentsCount: 0,
retrievedSegments: []int{0, 1},
retrievedParitySegments: []int{},
segmentationLayerV1ShouldSucceed: true,
segmentationLayerV2ShouldSucceed: true,
},
{
name: "all segments retrieved out of order",
segmentsCount: 2,
expectedParitySegmentsCount: 0,
retrievedSegments: []int{1, 0},
retrievedParitySegments: []int{},
segmentationLayerV1ShouldSucceed: true,
segmentationLayerV2ShouldSucceed: true,
},
{
name: "all segments&parity retrieved",
segmentsCount: 8,
expectedParitySegmentsCount: 1,
retrievedSegments: []int{0, 1, 2, 3, 4, 5, 6, 7, 8},
retrievedParitySegments: []int{8},
segmentationLayerV1ShouldSucceed: true,
segmentationLayerV2ShouldSucceed: true,
},
{
name: "all segments&parity retrieved out of order",
segmentsCount: 8,
expectedParitySegmentsCount: 1,
retrievedSegments: []int{8, 0, 7, 1, 6, 2, 5, 3, 4},
retrievedParitySegments: []int{8},
segmentationLayerV1ShouldSucceed: true,
segmentationLayerV2ShouldSucceed: true,
},
{
name: "no segments retrieved",
segmentsCount: 2,
expectedParitySegmentsCount: 0,
retrievedSegments: []int{},
retrievedParitySegments: []int{},
segmentationLayerV1ShouldSucceed: false,
segmentationLayerV2ShouldSucceed: false,
},
{
name: "not all needed segments&parity retrieved",
segmentsCount: 8,
expectedParitySegmentsCount: 1,
retrievedSegments: []int{1, 2, 8},
retrievedParitySegments: []int{8},
segmentationLayerV1ShouldSucceed: false,
segmentationLayerV2ShouldSucceed: false,
},
{
name: "segments&parity retrieved",
segmentsCount: 8,
expectedParitySegmentsCount: 1,
retrievedSegments: []int{1, 2, 3, 4, 5, 6, 7, 8},
retrievedParitySegments: []int{8},
segmentationLayerV1ShouldSucceed: false,
segmentationLayerV2ShouldSucceed: true, // succeed even though one segment is missing, thank you reedsolomon
},
{
name: "segments&parity retrieved out of order",
segmentsCount: 16,
expectedParitySegmentsCount: 2,
retrievedSegments: []int{17, 0, 16, 1, 15, 2, 14, 3, 13, 4, 12, 5, 11, 6, 10, 7},
retrievedParitySegments: []int{16, 17},
segmentationLayerV1ShouldSucceed: false,
segmentationLayerV2ShouldSucceed: true, // succeed even though two segments are missing, thank you reedsolomon
},
}
for _, version := range []string{"V1", "V2"} {
for _, tc := range testCases {
s.Run(fmt.Sprintf("%s %s", version, tc.name), func() {
segmentedMessages, err := segmentMessage(&types.NewMessage{Payload: s.testPayload}, int(math.Ceil(float64(len(s.testPayload))/float64(tc.segmentsCount))))
s.Require().NoError(err)
s.Require().Len(segmentedMessages, tc.segmentsCount+tc.expectedParitySegmentsCount)
message := &protocol.StatusMessage{TransportLayer: protocol.TransportLayer{
SigPubKey: &s.sender.identity.PublicKey,
}}
messageRecreated := false
handledSegments := []int{}
for i, segmentIndex := range tc.retrievedSegments {
s.T().Log("i=", i, "segmentIndex=", segmentIndex)
message.TransportLayer.Payload = segmentedMessages[segmentIndex].Payload
if version == "V1" {
err = s.sender.handleSegmentationLayerV1(message)
// V1 is unable to handle parity segment
if slices.Contains(tc.retrievedParitySegments, segmentIndex) {
if len(handledSegments) >= tc.segmentsCount {
s.Require().ErrorIs(err, ErrMessageSegmentsAlreadyCompleted)
} else {
s.Require().ErrorIs(err, ErrMessageSegmentsInvalidCount)
}
continue
}
} else {
err = s.sender.handleSegmentationLayerV2(message)
}
handledSegments = append(handledSegments, segmentIndex)
if len(handledSegments) < tc.segmentsCount {
s.Require().ErrorIs(err, ErrMessageSegmentsIncomplete)
} else if len(handledSegments) == tc.segmentsCount {
s.Require().NoError(err)
s.Require().ElementsMatch(s.testPayload, message.TransportLayer.Payload)
messageRecreated = true
} else {
s.Require().ErrorIs(err, ErrMessageSegmentsAlreadyCompleted)
}
}
if version == "V1" {
s.Require().Equal(tc.segmentationLayerV1ShouldSucceed, messageRecreated)
} else {
s.Require().Equal(tc.segmentationLayerV2ShouldSucceed, messageRecreated)
}
})
}
}
}

View File

@ -1,16 +1,13 @@
package common package common
import ( import (
"bytes"
"context" "context"
"crypto/ecdsa" "crypto/ecdsa"
"database/sql" "database/sql"
"math"
"sync" "sync"
"time" "time"
"github.com/golang/protobuf/proto" "github.com/golang/protobuf/proto"
"github.com/jinzhu/copier"
"github.com/pkg/errors" "github.com/pkg/errors"
datasyncnode "github.com/status-im/mvds/node" datasyncnode "github.com/status-im/mvds/node"
datasyncproto "github.com/status-im/mvds/protobuf" datasyncproto "github.com/status-im/mvds/protobuf"
@ -897,7 +894,7 @@ func (s *MessageSender) handleMessage(wakuMessage *types.Message) (*handleMessag
return nil, err return nil, err
} }
err = s.handleSegmentationLayer(message) err = s.handleSegmentationLayerV2(message)
if err != nil { if err != nil {
hlogger.Debug("failed to handle segmentation layer message", zap.Error(err)) hlogger.Debug("failed to handle segmentation layer message", zap.Error(err))
@ -1281,151 +1278,3 @@ func (s *MessageSender) GetCurrentKeyForGroup(groupID []byte) (*encryption.HashR
func (s *MessageSender) GetKeysForGroup(groupID []byte) ([]*encryption.HashRatchetKeyCompatibility, error) { func (s *MessageSender) GetKeysForGroup(groupID []byte) ([]*encryption.HashRatchetKeyCompatibility, error) {
return s.protocol.GetKeysForGroup(groupID) return s.protocol.GetKeysForGroup(groupID)
} }
// Segments message into smaller chunks if the size exceeds the maximum allowed
func segmentMessage(newMessage *types.NewMessage, maxSegmentSize int) ([]*types.NewMessage, error) {
if len(newMessage.Payload) <= maxSegmentSize {
return []*types.NewMessage{newMessage}, nil
}
createSegment := func(chunkPayload []byte) (*types.NewMessage, error) {
copy := &types.NewMessage{}
err := copier.Copy(copy, newMessage)
if err != nil {
return nil, err
}
copy.Payload = chunkPayload
copy.PowTarget = calculatePoW(chunkPayload)
return copy, nil
}
entireMessageHash := crypto.Keccak256(newMessage.Payload)
payloadSize := len(newMessage.Payload)
segmentsCount := int(math.Ceil(float64(payloadSize) / float64(maxSegmentSize)))
var segmentMessages []*types.NewMessage
for start, index := 0, 0; start < payloadSize; start += maxSegmentSize {
end := start + maxSegmentSize
if end > payloadSize {
end = payloadSize
}
chunk := newMessage.Payload[start:end]
segmentMessageProto := &protobuf.SegmentMessage{
EntireMessageHash: entireMessageHash,
Index: uint32(index),
SegmentsCount: uint32(segmentsCount),
Payload: chunk,
}
chunkPayload, err := proto.Marshal(segmentMessageProto)
if err != nil {
return nil, err
}
segmentMessage, err := createSegment(chunkPayload)
if err != nil {
return nil, err
}
segmentMessages = append(segmentMessages, segmentMessage)
index++
}
return segmentMessages, nil
}
func (s *MessageSender) segmentMessage(newMessage *types.NewMessage) ([]*types.NewMessage, error) {
// We set the max message size to 3/4 of the allowed message size, to leave
// room for segment message metadata.
newMessages, err := segmentMessage(newMessage, int(s.transport.MaxMessageSize()/4*3))
s.logger.Debug("message segmented", zap.Int("segments", len(newMessages)))
return newMessages, err
}
var ErrMessageSegmentsIncomplete = errors.New("message segments incomplete")
var ErrMessageSegmentsAlreadyCompleted = errors.New("message segments already completed")
var ErrMessageSegmentsInvalidCount = errors.New("invalid segments count")
var ErrMessageSegmentsHashMismatch = errors.New("hash of entire payload does not match")
func (s *MessageSender) handleSegmentationLayer(message *v1protocol.StatusMessage) error {
logger := s.logger.With(zap.String("site", "handleSegmentationLayer"))
hlogger := logger.With(zap.String("hash", types.HexBytes(message.TransportLayer.Hash).String()))
var segmentMessage protobuf.SegmentMessage
err := proto.Unmarshal(message.TransportLayer.Payload, &segmentMessage)
if err != nil {
return errors.Wrap(err, "failed to unmarshal SegmentMessage")
}
hlogger.Debug("handling message segment", zap.String("EntireMessageHash", types.HexBytes(segmentMessage.EntireMessageHash).String()),
zap.Uint32("Index", segmentMessage.Index), zap.Uint32("SegmentsCount", segmentMessage.SegmentsCount))
alreadyCompleted, err := s.persistence.IsMessageAlreadyCompleted(segmentMessage.EntireMessageHash)
if err != nil {
return err
}
if alreadyCompleted {
return ErrMessageSegmentsAlreadyCompleted
}
if segmentMessage.SegmentsCount < 2 {
return ErrMessageSegmentsInvalidCount
}
err = s.persistence.SaveMessageSegment(&segmentMessage, message.TransportLayer.SigPubKey, time.Now().Unix())
if err != nil {
return err
}
segments, err := s.persistence.GetMessageSegments(segmentMessage.EntireMessageHash, message.TransportLayer.SigPubKey)
if err != nil {
return err
}
if len(segments) != int(segmentMessage.SegmentsCount) {
return ErrMessageSegmentsIncomplete
}
// Combine payload
var entirePayload bytes.Buffer
for _, segment := range segments {
_, err := entirePayload.Write(segment.Payload)
if err != nil {
return errors.Wrap(err, "failed to write segment payload")
}
}
// Sanity check
entirePayloadHash := crypto.Keccak256(entirePayload.Bytes())
if !bytes.Equal(entirePayloadHash, segmentMessage.EntireMessageHash) {
return ErrMessageSegmentsHashMismatch
}
err = s.persistence.CompleteMessageSegments(segmentMessage.EntireMessageHash, message.TransportLayer.SigPubKey, time.Now().Unix())
if err != nil {
return err
}
message.TransportLayer.Payload = entirePayload.Bytes()
return nil
}
func (s *MessageSender) CleanupSegments() error {
weekAgo := time.Now().AddDate(0, 0, -7).Unix()
monthAgo := time.Now().AddDate(0, -1, 0).Unix()
err := s.persistence.RemoveMessageSegmentsOlderThan(weekAgo)
if err != nil {
return err
}
err = s.persistence.RemoveMessageSegmentsCompletedOlderThan(monthAgo)
if err != nil {
return err
}
return nil
}

View File

@ -347,33 +347,46 @@ func (db *RawMessagesPersistence) IsMessageAlreadyCompleted(hash []byte) (bool,
return alreadyCompleted > 0, nil return alreadyCompleted > 0, nil
} }
func (db *RawMessagesPersistence) SaveMessageSegment(segment *protobuf.SegmentMessage, sigPubKey *ecdsa.PublicKey, timestamp int64) error { func (db *RawMessagesPersistence) SaveMessageSegment(segment *SegmentMessage, sigPubKey *ecdsa.PublicKey, timestamp int64) error {
sigPubKeyBlob := crypto.CompressPubkey(sigPubKey) sigPubKeyBlob := crypto.CompressPubkey(sigPubKey)
_, err := db.db.Exec("INSERT INTO message_segments (hash, segment_index, segments_count, sig_pub_key, payload, timestamp) VALUES (?, ?, ?, ?, ?, ?)", _, err := db.db.Exec("INSERT INTO message_segments (hash, segment_index, segments_count, parity_segment_index, parity_segments_count, sig_pub_key, payload, timestamp) VALUES (?, ?, ?, ?, ?, ?, ?, ?)",
segment.EntireMessageHash, segment.Index, segment.SegmentsCount, sigPubKeyBlob, segment.Payload, timestamp) segment.EntireMessageHash, segment.Index, segment.SegmentsCount, segment.ParitySegmentIndex, segment.ParitySegmentsCount, sigPubKeyBlob, segment.Payload, timestamp)
return err return err
} }
// Get ordered message segments for given hash // Get ordered message segments for given hash
func (db *RawMessagesPersistence) GetMessageSegments(hash []byte, sigPubKey *ecdsa.PublicKey) ([]*protobuf.SegmentMessage, error) { func (db *RawMessagesPersistence) GetMessageSegments(hash []byte, sigPubKey *ecdsa.PublicKey) ([]*SegmentMessage, error) {
sigPubKeyBlob := crypto.CompressPubkey(sigPubKey) sigPubKeyBlob := crypto.CompressPubkey(sigPubKey)
rows, err := db.db.Query("SELECT hash, segment_index, segments_count, payload FROM message_segments WHERE hash = ? AND sig_pub_key = ? ORDER BY segment_index", hash, sigPubKeyBlob) rows, err := db.db.Query(`
SELECT
hash, segment_index, segments_count, parity_segment_index, parity_segments_count, payload
FROM
message_segments
WHERE
hash = ? AND sig_pub_key = ?
ORDER BY
(segments_count = 0) ASC, -- Prioritize segments_count > 0
segment_index ASC,
parity_segment_index ASC`,
hash, sigPubKeyBlob)
if err != nil { if err != nil {
return nil, err return nil, err
} }
defer rows.Close() defer rows.Close()
var segments []*protobuf.SegmentMessage var segments []*SegmentMessage
for rows.Next() { for rows.Next() {
var segment protobuf.SegmentMessage segment := &SegmentMessage{
err := rows.Scan(&segment.EntireMessageHash, &segment.Index, &segment.SegmentsCount, &segment.Payload) SegmentMessage: &protobuf.SegmentMessage{},
}
err := rows.Scan(&segment.EntireMessageHash, &segment.Index, &segment.SegmentsCount, &segment.ParitySegmentIndex, &segment.ParitySegmentsCount, &segment.Payload)
if err != nil { if err != nil {
return nil, err return nil, err
} }
segments = append(segments, &segment) segments = append(segments, segment)
} }
err = rows.Err() err = rows.Err()
if err != nil { if err != nil {

View File

@ -134,6 +134,7 @@
// 1711389881_add_profile_showcase_community_grant.up.sql (86B) // 1711389881_add_profile_showcase_community_grant.up.sql (86B)
// 1711937186_add_contact_customization_color.up.sql (172B) // 1711937186_add_contact_customization_color.up.sql (172B)
// 1712745141_hash_ratchet_encrypted_messages_key_id.up.sql (111B) // 1712745141_hash_ratchet_encrypted_messages_key_id.up.sql (111B)
// 1712905223_add_parity_to_message_segments.up.sql (792B)
// README.md (554B) // README.md (554B)
// doc.go (870B) // doc.go (870B)
@ -2883,6 +2884,26 @@ func _1712745141_hash_ratchet_encrypted_messages_key_idUpSql() (*asset, error) {
return a, nil return a, nil
} }
var __1712905223_add_parity_to_message_segmentsUpSql = []byte("\x1f\x8b\x08\x00\x00\x00\x00\x00\x00\xff\xac\x92\x3d\x6b\xc3\x30\x10\x86\x77\xfd\x8a\x1b\x13\xd0\x90\x3d\x93\xec\x5e\x8a\xa9\x2c\x99\x8b\x3a\x64\x12\x6e\x2d\x1c\xd3\xf8\x83\x4a\x81\xfa\xdf\x97\x04\x53\x1a\x23\xf7\x03\xb2\xde\x3d\xd2\x3d\xbc\x77\x42\x1a\x24\x30\x22\x91\x08\xad\xf3\xbe\xac\x9d\xf5\xae\x6e\x5d\x17\x3c\x10\x2a\x91\x23\x18\x0d\xfd\xa9\xb2\xf3\xf6\x96\xb1\x94\x50\x18\x5c\x7a\xbe\x62\x00\x00\xc7\xd2\x1f\x21\x91\x3a\x01\xa5\x0d\xa8\x67\x29\xf9\xb5\x3e\x61\xb6\xe9\x2a\xf7\x01\x99\x32\xf8\x88\x14\x67\xbc\x7d\xed\xcf\x5d\x58\x80\x86\x72\x3c\xf5\x65\x15\x9d\xd1\xd4\x76\x38\xbf\xd8\x37\x37\xc6\xda\xa1\x69\x9d\x0f\x65\x3b\x2c\xfe\xfc\xde\x84\xd1\xfe\xc5\xf4\x16\xfd\x59\xb8\xa0\x2c\x17\x74\x80\x27\x3c\xc0\xea\x12\x0f\xff\x2e\xca\x6f\x93\xe1\xb3\x10\x78\xd4\x6a\x5e\x9d\xe0\x35\x68\x05\xa9\x56\x3b\x99\xa5\x06\x08\x0b\x29\x52\x64\xeb\x2d\x63\x99\xda\x23\x99\x8b\xa0\x8e\x2c\x6e\x92\xfa\xcd\xe3\x9a\xfb\x4c\xfe\x2b\xd3\xff\x89\xb2\x3d\x4a\x4c\x0d\xdc\x67\xf2\x86\xc3\x86\xed\x48\xe7\x4b\x87\xfb\x40\xba\x98\xce\x36\x4e\x7c\x06\x00\x00\xff\xff\xfd\x2e\x92\x5d\x18\x03\x00\x00")
func _1712905223_add_parity_to_message_segmentsUpSqlBytes() ([]byte, error) {
return bindataRead(
__1712905223_add_parity_to_message_segmentsUpSql,
"1712905223_add_parity_to_message_segments.up.sql",
)
}
func _1712905223_add_parity_to_message_segmentsUpSql() (*asset, error) {
bytes, err := _1712905223_add_parity_to_message_segmentsUpSqlBytes()
if err != nil {
return nil, err
}
info := bindataFileInfo{name: "1712905223_add_parity_to_message_segments.up.sql", size: 792, mode: os.FileMode(0644), modTime: time.Unix(1700000000, 0)}
a := &asset{bytes: bytes, info: info, digest: [32]uint8{0x9, 0x78, 0x5e, 0x84, 0x2b, 0xf9, 0x52, 0x77, 0x7, 0x6c, 0xb6, 0x76, 0x6d, 0x59, 0xd2, 0x1f, 0x6c, 0xe6, 0xd, 0x86, 0x85, 0xeb, 0x34, 0x95, 0x6e, 0xa, 0xa9, 0xd8, 0x3b, 0x7a, 0xd, 0x1a}}
return a, nil
}
var _readmeMd = []byte("\x1f\x8b\x08\x00\x00\x00\x00\x00\x00\xff\x54\x91\xc1\xce\xd3\x30\x10\x84\xef\x7e\x8a\x91\x7a\x01\xa9\x2a\x8f\xc0\x0d\x71\x82\x03\x48\x1c\xc9\x36\x9e\x36\x96\x1c\x6f\xf0\xae\x93\xe6\xed\x91\xa3\xc2\xdf\xff\x66\xed\xd8\x33\xdf\x78\x4f\xa7\x13\xbe\xea\x06\x57\x6c\x35\x39\x31\xa7\x7b\x15\x4f\x5a\xec\x73\x08\xbf\x08\x2d\x79\x7f\x4a\x43\x5b\x86\x17\xfd\x8c\x21\xea\x56\x5e\x47\x90\x4a\x14\x75\x48\xde\x64\x37\x2c\x6a\x96\xae\x99\x48\x05\xf6\x27\x77\x13\xad\x08\xae\x8a\x51\xe7\x25\xf3\xf1\xa9\x9f\xf9\x58\x58\x2c\xad\xbc\xe0\x8b\x56\xf0\x21\x5d\xeb\x4c\x95\xb3\xae\x84\x60\xd4\xdc\xe6\x82\x5d\x1b\x36\x6d\x39\x62\x92\xf5\xb8\x11\xdb\x92\xd3\x28\xce\xe0\x13\xe1\x72\xcd\x3c\x63\xd4\x65\x87\xae\xac\xe8\xc3\x28\x2e\x67\x44\x66\x3a\x21\x25\xa2\x72\xac\x14\x67\xbc\x84\x9f\x53\x32\x8c\x52\x70\x25\x56\xd6\xfd\x8d\x05\x37\xad\x30\x9d\x9f\xa6\x86\x0f\xcd\x58\x7f\xcf\x34\x93\x3b\xed\x90\x9f\xa4\x1f\xcf\x30\x85\x4d\x07\x58\xaf\x7f\x25\xc4\x9d\xf3\x72\x64\x84\xd0\x7f\xf9\x9b\x3a\x2d\x84\xef\x85\x48\x66\x8d\xd8\x88\x9b\x8c\x8c\x98\x5b\xf6\x74\x14\x4e\x33\x0d\xc9\xe0\x93\x38\xda\x12\xc5\x69\xbd\xe4\xf0\x2e\x7a\x78\x07\x1c\xfe\x13\x9f\x91\x29\x31\x95\x7b\x7f\x62\x59\x37\xb4\xe5\x5e\x25\xfe\x33\xee\xd5\x53\x71\xd6\xda\x3a\xd8\xcb\xde\x2e\xf8\xa1\x90\x55\x53\x0c\xc7\xaa\x0d\xe9\x76\x14\x29\x1c\x7b\x68\xdd\x2f\xe1\x6f\x00\x00\x00\xff\xff\x3c\x0a\xc2\xfe\x2a\x02\x00\x00") var _readmeMd = []byte("\x1f\x8b\x08\x00\x00\x00\x00\x00\x00\xff\x54\x91\xc1\xce\xd3\x30\x10\x84\xef\x7e\x8a\x91\x7a\x01\xa9\x2a\x8f\xc0\x0d\x71\x82\x03\x48\x1c\xc9\x36\x9e\x36\x96\x1c\x6f\xf0\xae\x93\xe6\xed\x91\xa3\xc2\xdf\xff\x66\xed\xd8\x33\xdf\x78\x4f\xa7\x13\xbe\xea\x06\x57\x6c\x35\x39\x31\xa7\x7b\x15\x4f\x5a\xec\x73\x08\xbf\x08\x2d\x79\x7f\x4a\x43\x5b\x86\x17\xfd\x8c\x21\xea\x56\x5e\x47\x90\x4a\x14\x75\x48\xde\x64\x37\x2c\x6a\x96\xae\x99\x48\x05\xf6\x27\x77\x13\xad\x08\xae\x8a\x51\xe7\x25\xf3\xf1\xa9\x9f\xf9\x58\x58\x2c\xad\xbc\xe0\x8b\x56\xf0\x21\x5d\xeb\x4c\x95\xb3\xae\x84\x60\xd4\xdc\xe6\x82\x5d\x1b\x36\x6d\x39\x62\x92\xf5\xb8\x11\xdb\x92\xd3\x28\xce\xe0\x13\xe1\x72\xcd\x3c\x63\xd4\x65\x87\xae\xac\xe8\xc3\x28\x2e\x67\x44\x66\x3a\x21\x25\xa2\x72\xac\x14\x67\xbc\x84\x9f\x53\x32\x8c\x52\x70\x25\x56\xd6\xfd\x8d\x05\x37\xad\x30\x9d\x9f\xa6\x86\x0f\xcd\x58\x7f\xcf\x34\x93\x3b\xed\x90\x9f\xa4\x1f\xcf\x30\x85\x4d\x07\x58\xaf\x7f\x25\xc4\x9d\xf3\x72\x64\x84\xd0\x7f\xf9\x9b\x3a\x2d\x84\xef\x85\x48\x66\x8d\xd8\x88\x9b\x8c\x8c\x98\x5b\xf6\x74\x14\x4e\x33\x0d\xc9\xe0\x93\x38\xda\x12\xc5\x69\xbd\xe4\xf0\x2e\x7a\x78\x07\x1c\xfe\x13\x9f\x91\x29\x31\x95\x7b\x7f\x62\x59\x37\xb4\xe5\x5e\x25\xfe\x33\xee\xd5\x53\x71\xd6\xda\x3a\xd8\xcb\xde\x2e\xf8\xa1\x90\x55\x53\x0c\xc7\xaa\x0d\xe9\x76\x14\x29\x1c\x7b\x68\xdd\x2f\xe1\x6f\x00\x00\x00\xff\xff\x3c\x0a\xc2\xfe\x2a\x02\x00\x00")
func readmeMdBytes() ([]byte, error) { func readmeMdBytes() ([]byte, error) {
@ -3148,6 +3169,7 @@ var _bindata = map[string]func() (*asset, error){
"1711389881_add_profile_showcase_community_grant.up.sql": _1711389881_add_profile_showcase_community_grantUpSql, "1711389881_add_profile_showcase_community_grant.up.sql": _1711389881_add_profile_showcase_community_grantUpSql,
"1711937186_add_contact_customization_color.up.sql": _1711937186_add_contact_customization_colorUpSql, "1711937186_add_contact_customization_color.up.sql": _1711937186_add_contact_customization_colorUpSql,
"1712745141_hash_ratchet_encrypted_messages_key_id.up.sql": _1712745141_hash_ratchet_encrypted_messages_key_idUpSql, "1712745141_hash_ratchet_encrypted_messages_key_id.up.sql": _1712745141_hash_ratchet_encrypted_messages_key_idUpSql,
"1712905223_add_parity_to_message_segments.up.sql": _1712905223_add_parity_to_message_segmentsUpSql,
"README.md": readmeMd, "README.md": readmeMd,
"doc.go": docGo, "doc.go": docGo,
} }
@ -3332,6 +3354,7 @@ var _bintree = &bintree{nil, map[string]*bintree{
"1711389881_add_profile_showcase_community_grant.up.sql": {_1711389881_add_profile_showcase_community_grantUpSql, map[string]*bintree{}}, "1711389881_add_profile_showcase_community_grant.up.sql": {_1711389881_add_profile_showcase_community_grantUpSql, map[string]*bintree{}},
"1711937186_add_contact_customization_color.up.sql": {_1711937186_add_contact_customization_colorUpSql, map[string]*bintree{}}, "1711937186_add_contact_customization_color.up.sql": {_1711937186_add_contact_customization_colorUpSql, map[string]*bintree{}},
"1712745141_hash_ratchet_encrypted_messages_key_id.up.sql": {_1712745141_hash_ratchet_encrypted_messages_key_idUpSql, map[string]*bintree{}}, "1712745141_hash_ratchet_encrypted_messages_key_id.up.sql": {_1712745141_hash_ratchet_encrypted_messages_key_idUpSql, map[string]*bintree{}},
"1712905223_add_parity_to_message_segments.up.sql": {_1712905223_add_parity_to_message_segmentsUpSql, map[string]*bintree{}},
"README.md": {readmeMd, map[string]*bintree{}}, "README.md": {readmeMd, map[string]*bintree{}},
"doc.go": {docGo, map[string]*bintree{}}, "doc.go": {docGo, map[string]*bintree{}},
}} }}

View File

@ -0,0 +1,19 @@
ALTER TABLE message_segments RENAME TO old_message_segments;
CREATE TABLE message_segments (
hash BLOB NOT NULL,
segment_index INTEGER NOT NULL,
segments_count INTEGER NOT NULL,
payload BLOB NOT NULL,
sig_pub_key BLOB NOT NULL,
timestamp INTEGER NOT NULL,
parity_segment_index INTEGER NOT NULL,
parity_segments_count INTEGER NOT NULL,
PRIMARY KEY (hash, sig_pub_key, segment_index, segments_count, parity_segment_index, parity_segments_count) ON CONFLICT REPLACE
);
INSERT INTO message_segments (hash, segment_index, segments_count, payload, sig_pub_key, timestamp, parity_segment_index, parity_segments_count)
SELECT hash, segment_index, segments_count, payload, sig_pub_key, timestamp, 0, 0
FROM old_message_segments;
DROP TABLE old_message_segments;

View File

@ -33,6 +33,10 @@ type SegmentMessage struct {
SegmentsCount uint32 `protobuf:"varint,3,opt,name=segments_count,json=segmentsCount,proto3" json:"segments_count,omitempty"` SegmentsCount uint32 `protobuf:"varint,3,opt,name=segments_count,json=segmentsCount,proto3" json:"segments_count,omitempty"`
// The payload data for this particular segment // The payload data for this particular segment
Payload []byte `protobuf:"bytes,4,opt,name=payload,proto3" json:"payload,omitempty"` Payload []byte `protobuf:"bytes,4,opt,name=payload,proto3" json:"payload,omitempty"`
// Index of this parity segment
ParitySegmentIndex uint32 `protobuf:"varint,5,opt,name=parity_segment_index,json=paritySegmentIndex,proto3" json:"parity_segment_index,omitempty"`
// Total number of parity segments
ParitySegmentsCount uint32 `protobuf:"varint,6,opt,name=parity_segments_count,json=paritySegmentsCount,proto3" json:"parity_segments_count,omitempty"`
} }
func (x *SegmentMessage) Reset() { func (x *SegmentMessage) Reset() {
@ -95,12 +99,26 @@ func (x *SegmentMessage) GetPayload() []byte {
return nil return nil
} }
func (x *SegmentMessage) GetParitySegmentIndex() uint32 {
if x != nil {
return x.ParitySegmentIndex
}
return 0
}
func (x *SegmentMessage) GetParitySegmentsCount() uint32 {
if x != nil {
return x.ParitySegmentsCount
}
return 0
}
var File_segment_message_proto protoreflect.FileDescriptor var File_segment_message_proto protoreflect.FileDescriptor
var file_segment_message_proto_rawDesc = []byte{ var file_segment_message_proto_rawDesc = []byte{
0x0a, 0x15, 0x73, 0x65, 0x67, 0x6d, 0x65, 0x6e, 0x74, 0x5f, 0x6d, 0x65, 0x73, 0x73, 0x61, 0x67, 0x0a, 0x15, 0x73, 0x65, 0x67, 0x6d, 0x65, 0x6e, 0x74, 0x5f, 0x6d, 0x65, 0x73, 0x73, 0x61, 0x67,
0x65, 0x2e, 0x70, 0x72, 0x6f, 0x74, 0x6f, 0x12, 0x08, 0x70, 0x72, 0x6f, 0x74, 0x6f, 0x62, 0x75, 0x65, 0x2e, 0x70, 0x72, 0x6f, 0x74, 0x6f, 0x12, 0x08, 0x70, 0x72, 0x6f, 0x74, 0x6f, 0x62, 0x75,
0x66, 0x22, 0x97, 0x01, 0x0a, 0x0e, 0x53, 0x65, 0x67, 0x6d, 0x65, 0x6e, 0x74, 0x4d, 0x65, 0x73, 0x66, 0x22, 0xfd, 0x01, 0x0a, 0x0e, 0x53, 0x65, 0x67, 0x6d, 0x65, 0x6e, 0x74, 0x4d, 0x65, 0x73,
0x73, 0x61, 0x67, 0x65, 0x12, 0x2e, 0x0a, 0x13, 0x65, 0x6e, 0x74, 0x69, 0x72, 0x65, 0x5f, 0x6d, 0x73, 0x61, 0x67, 0x65, 0x12, 0x2e, 0x0a, 0x13, 0x65, 0x6e, 0x74, 0x69, 0x72, 0x65, 0x5f, 0x6d,
0x65, 0x73, 0x73, 0x61, 0x67, 0x65, 0x5f, 0x68, 0x61, 0x73, 0x68, 0x18, 0x01, 0x20, 0x01, 0x28, 0x65, 0x73, 0x73, 0x61, 0x67, 0x65, 0x5f, 0x68, 0x61, 0x73, 0x68, 0x18, 0x01, 0x20, 0x01, 0x28,
0x0c, 0x52, 0x11, 0x65, 0x6e, 0x74, 0x69, 0x72, 0x65, 0x4d, 0x65, 0x73, 0x73, 0x61, 0x67, 0x65, 0x0c, 0x52, 0x11, 0x65, 0x6e, 0x74, 0x69, 0x72, 0x65, 0x4d, 0x65, 0x73, 0x73, 0x61, 0x67, 0x65,
@ -109,9 +127,15 @@ var file_segment_message_proto_rawDesc = []byte{
0x67, 0x6d, 0x65, 0x6e, 0x74, 0x73, 0x5f, 0x63, 0x6f, 0x75, 0x6e, 0x74, 0x18, 0x03, 0x20, 0x01, 0x67, 0x6d, 0x65, 0x6e, 0x74, 0x73, 0x5f, 0x63, 0x6f, 0x75, 0x6e, 0x74, 0x18, 0x03, 0x20, 0x01,
0x28, 0x0d, 0x52, 0x0d, 0x73, 0x65, 0x67, 0x6d, 0x65, 0x6e, 0x74, 0x73, 0x43, 0x6f, 0x75, 0x6e, 0x28, 0x0d, 0x52, 0x0d, 0x73, 0x65, 0x67, 0x6d, 0x65, 0x6e, 0x74, 0x73, 0x43, 0x6f, 0x75, 0x6e,
0x74, 0x12, 0x18, 0x0a, 0x07, 0x70, 0x61, 0x79, 0x6c, 0x6f, 0x61, 0x64, 0x18, 0x04, 0x20, 0x01, 0x74, 0x12, 0x18, 0x0a, 0x07, 0x70, 0x61, 0x79, 0x6c, 0x6f, 0x61, 0x64, 0x18, 0x04, 0x20, 0x01,
0x28, 0x0c, 0x52, 0x07, 0x70, 0x61, 0x79, 0x6c, 0x6f, 0x61, 0x64, 0x42, 0x0d, 0x5a, 0x0b, 0x2e, 0x28, 0x0c, 0x52, 0x07, 0x70, 0x61, 0x79, 0x6c, 0x6f, 0x61, 0x64, 0x12, 0x30, 0x0a, 0x14, 0x70,
0x2f, 0x3b, 0x70, 0x72, 0x6f, 0x74, 0x6f, 0x62, 0x75, 0x66, 0x62, 0x06, 0x70, 0x72, 0x6f, 0x74, 0x61, 0x72, 0x69, 0x74, 0x79, 0x5f, 0x73, 0x65, 0x67, 0x6d, 0x65, 0x6e, 0x74, 0x5f, 0x69, 0x6e,
0x6f, 0x33, 0x64, 0x65, 0x78, 0x18, 0x05, 0x20, 0x01, 0x28, 0x0d, 0x52, 0x12, 0x70, 0x61, 0x72, 0x69, 0x74,
0x79, 0x53, 0x65, 0x67, 0x6d, 0x65, 0x6e, 0x74, 0x49, 0x6e, 0x64, 0x65, 0x78, 0x12, 0x32, 0x0a,
0x15, 0x70, 0x61, 0x72, 0x69, 0x74, 0x79, 0x5f, 0x73, 0x65, 0x67, 0x6d, 0x65, 0x6e, 0x74, 0x73,
0x5f, 0x63, 0x6f, 0x75, 0x6e, 0x74, 0x18, 0x06, 0x20, 0x01, 0x28, 0x0d, 0x52, 0x13, 0x70, 0x61,
0x72, 0x69, 0x74, 0x79, 0x53, 0x65, 0x67, 0x6d, 0x65, 0x6e, 0x74, 0x73, 0x43, 0x6f, 0x75, 0x6e,
0x74, 0x42, 0x0d, 0x5a, 0x0b, 0x2e, 0x2f, 0x3b, 0x70, 0x72, 0x6f, 0x74, 0x6f, 0x62, 0x75, 0x66,
0x62, 0x06, 0x70, 0x72, 0x6f, 0x74, 0x6f, 0x33,
} }
var ( var (

View File

@ -12,4 +12,8 @@ message SegmentMessage {
uint32 segments_count = 3; uint32 segments_count = 3;
// The payload data for this particular segment // The payload data for this particular segment
bytes payload = 4; bytes payload = 4;
// Index of this parity segment
uint32 parity_segment_index = 5;
// Total number of parity segments
uint32 parity_segments_count = 6;
} }