954 lines
26 KiB
Go
954 lines
26 KiB
Go
// Copyright 2017 The go-ethereum Authors
|
|
// This file is part of the go-ethereum library.
|
|
//
|
|
// The go-ethereum library is free software: you can redistribute it and/or modify
|
|
// it under the terms of the GNU Lesser General Public License as published by
|
|
// the Free Software Foundation, either version 3 of the License, or
|
|
// (at your option) any later version.
|
|
//
|
|
// The go-ethereum library is distributed in the hope that it will be useful,
|
|
// but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
// GNU Lesser General Public License for more details.
|
|
//
|
|
// You should have received a copy of the GNU Lesser General Public License
|
|
// along with the go-ethereum library. If not, see <http://www.gnu.org/licenses/>.
|
|
|
|
package mailserver
|
|
|
|
import (
|
|
"crypto/ecdsa"
|
|
"encoding/binary"
|
|
"errors"
|
|
"fmt"
|
|
"math/rand"
|
|
"sync"
|
|
"time"
|
|
|
|
prom "github.com/prometheus/client_golang/prometheus"
|
|
"go.uber.org/zap"
|
|
|
|
"github.com/ethereum/go-ethereum/common"
|
|
"github.com/ethereum/go-ethereum/rlp"
|
|
gocommon "github.com/status-im/status-go/common"
|
|
gethbridge "github.com/status-im/status-go/eth-node/bridge/geth"
|
|
"github.com/status-im/status-go/eth-node/crypto"
|
|
"github.com/status-im/status-go/eth-node/types"
|
|
"github.com/status-im/status-go/logutils"
|
|
"github.com/status-im/status-go/params"
|
|
"github.com/status-im/status-go/waku"
|
|
wakucommon "github.com/status-im/status-go/waku/common"
|
|
)
|
|
|
|
const (
|
|
maxQueryRange = 24 * time.Hour
|
|
maxQueryLimit = 1000
|
|
// When we default the upper limit, we want to extend the range a bit
|
|
// to accommodate for envelopes with slightly higher timestamp, in seconds
|
|
whisperTTLSafeThreshold = 60
|
|
)
|
|
|
|
var (
|
|
errDirectoryNotProvided = errors.New("data directory not provided")
|
|
errDecryptionMethodNotProvided = errors.New("decryption method is not provided")
|
|
)
|
|
|
|
const (
|
|
timestampLength = 4
|
|
requestLimitLength = 4
|
|
requestTimeRangeLength = timestampLength * 2
|
|
processRequestTimeout = time.Minute
|
|
)
|
|
|
|
type Config struct {
|
|
// DataDir points to a directory where mailserver's data is stored.
|
|
DataDir string
|
|
// Password is used to create a symmetric key to decrypt requests.
|
|
Password string
|
|
// AsymKey is an asymmetric key to decrypt requests.
|
|
AsymKey string
|
|
// MininumPoW is a minimum PoW for requests.
|
|
MinimumPoW float64
|
|
// RateLimit is a maximum number of requests per second from a peer.
|
|
RateLimit int
|
|
// DataRetention specifies a number of days an envelope should be stored for.
|
|
DataRetention int
|
|
PostgresEnabled bool
|
|
PostgresURI string
|
|
}
|
|
|
|
// --------------
|
|
// WakuMailServer
|
|
// --------------
|
|
|
|
type WakuMailServer struct {
|
|
ms *mailServer
|
|
shh *waku.Waku
|
|
minRequestPoW float64
|
|
|
|
symFilter *wakucommon.Filter
|
|
asymFilter *wakucommon.Filter
|
|
}
|
|
|
|
func (s *WakuMailServer) Init(waku *waku.Waku, cfg *params.WakuConfig) error {
|
|
s.shh = waku
|
|
s.minRequestPoW = cfg.MinimumPoW
|
|
|
|
config := Config{
|
|
DataDir: cfg.DataDir,
|
|
Password: cfg.MailServerPassword,
|
|
MinimumPoW: cfg.MinimumPoW,
|
|
DataRetention: cfg.MailServerDataRetention,
|
|
RateLimit: cfg.MailServerRateLimit,
|
|
PostgresEnabled: cfg.DatabaseConfig.PGConfig.Enabled,
|
|
PostgresURI: cfg.DatabaseConfig.PGConfig.URI,
|
|
}
|
|
var err error
|
|
s.ms, err = newMailServer(
|
|
config,
|
|
&wakuAdapter{},
|
|
&wakuService{Waku: waku},
|
|
)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
if err := s.setupDecryptor(config.Password, config.AsymKey); err != nil {
|
|
return err
|
|
}
|
|
|
|
return nil
|
|
}
|
|
|
|
func (s *WakuMailServer) Close() {
|
|
s.ms.Close()
|
|
}
|
|
|
|
func (s *WakuMailServer) Archive(env *wakucommon.Envelope) {
|
|
s.ms.Archive(gethbridge.NewWakuEnvelope(env))
|
|
}
|
|
|
|
func (s *WakuMailServer) Deliver(peerID []byte, req wakucommon.MessagesRequest) {
|
|
s.ms.DeliverMail(types.BytesToHash(peerID), types.BytesToHash(req.ID), MessagesRequestPayload{
|
|
Lower: req.From,
|
|
Upper: req.To,
|
|
Bloom: req.Bloom,
|
|
Topics: req.Topics,
|
|
Limit: req.Limit,
|
|
Cursor: req.Cursor,
|
|
Batch: true,
|
|
})
|
|
}
|
|
|
|
// DEPRECATED; user Deliver instead
|
|
func (s *WakuMailServer) DeliverMail(peerID []byte, req *wakucommon.Envelope) {
|
|
payload, err := s.decodeRequest(peerID, req)
|
|
if err != nil {
|
|
deliveryFailuresCounter.WithLabelValues("validation").Inc()
|
|
logutils.ZapLogger().Error(
|
|
"[mailserver:DeliverMail] request failed validaton",
|
|
zap.Stringer("peerID", types.BytesToHash(peerID)),
|
|
zap.Stringer("requestID", req.Hash()),
|
|
zap.Error(err),
|
|
)
|
|
s.ms.sendHistoricMessageErrorResponse(types.BytesToHash(peerID), types.Hash(req.Hash()), err)
|
|
return
|
|
}
|
|
|
|
s.ms.DeliverMail(types.BytesToHash(peerID), types.Hash(req.Hash()), payload)
|
|
}
|
|
|
|
// bloomFromReceivedMessage for a given whisper.ReceivedMessage it extracts the
|
|
// used bloom filter.
|
|
func (s *WakuMailServer) bloomFromReceivedMessage(msg *wakucommon.ReceivedMessage) ([]byte, error) {
|
|
payloadSize := len(msg.Payload)
|
|
|
|
if payloadSize < 8 {
|
|
return nil, errors.New("Undersized p2p request")
|
|
} else if payloadSize == 8 {
|
|
return wakucommon.MakeFullNodeBloom(), nil
|
|
} else if payloadSize < 8+wakucommon.BloomFilterSize {
|
|
return nil, errors.New("Undersized bloom filter in p2p request")
|
|
}
|
|
|
|
return msg.Payload[8 : 8+wakucommon.BloomFilterSize], nil
|
|
}
|
|
|
|
func (s *WakuMailServer) decompositeRequest(peerID []byte, request *wakucommon.Envelope) (MessagesRequestPayload, error) {
|
|
var (
|
|
payload MessagesRequestPayload
|
|
err error
|
|
)
|
|
|
|
if s.minRequestPoW > 0.0 && request.PoW() < s.minRequestPoW {
|
|
return payload, fmt.Errorf("PoW() is too low")
|
|
}
|
|
|
|
decrypted := s.openEnvelope(request)
|
|
if decrypted == nil {
|
|
return payload, fmt.Errorf("failed to decrypt p2p request")
|
|
}
|
|
|
|
if err := checkMsgSignature(decrypted.Src, peerID); err != nil {
|
|
return payload, err
|
|
}
|
|
|
|
payload.Bloom, err = s.bloomFromReceivedMessage(decrypted)
|
|
if err != nil {
|
|
return payload, err
|
|
}
|
|
|
|
payload.Lower = binary.BigEndian.Uint32(decrypted.Payload[:4])
|
|
payload.Upper = binary.BigEndian.Uint32(decrypted.Payload[4:8])
|
|
|
|
if payload.Upper < payload.Lower {
|
|
err := fmt.Errorf("query range is invalid: from > to (%d > %d)", payload.Lower, payload.Upper)
|
|
return payload, err
|
|
}
|
|
|
|
lowerTime := time.Unix(int64(payload.Lower), 0)
|
|
upperTime := time.Unix(int64(payload.Upper), 0)
|
|
if upperTime.Sub(lowerTime) > maxQueryRange {
|
|
err := fmt.Errorf("query range too big for peer %s", string(peerID))
|
|
return payload, err
|
|
}
|
|
|
|
if len(decrypted.Payload) >= requestTimeRangeLength+wakucommon.BloomFilterSize+requestLimitLength {
|
|
payload.Limit = binary.BigEndian.Uint32(decrypted.Payload[requestTimeRangeLength+wakucommon.BloomFilterSize:])
|
|
}
|
|
|
|
if len(decrypted.Payload) == requestTimeRangeLength+wakucommon.BloomFilterSize+requestLimitLength+DBKeyLength {
|
|
payload.Cursor = decrypted.Payload[requestTimeRangeLength+wakucommon.BloomFilterSize+requestLimitLength:]
|
|
}
|
|
|
|
return payload, nil
|
|
}
|
|
|
|
func (s *WakuMailServer) setupDecryptor(password, asymKey string) error {
|
|
s.symFilter = nil
|
|
s.asymFilter = nil
|
|
|
|
if password != "" {
|
|
keyID, err := s.shh.AddSymKeyFromPassword(password)
|
|
if err != nil {
|
|
return fmt.Errorf("create symmetric key: %v", err)
|
|
}
|
|
|
|
symKey, err := s.shh.GetSymKey(keyID)
|
|
if err != nil {
|
|
return fmt.Errorf("save symmetric key: %v", err)
|
|
}
|
|
|
|
s.symFilter = &wakucommon.Filter{KeySym: symKey}
|
|
}
|
|
|
|
if asymKey != "" {
|
|
keyAsym, err := crypto.HexToECDSA(asymKey)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
s.asymFilter = &wakucommon.Filter{KeyAsym: keyAsym}
|
|
}
|
|
|
|
return nil
|
|
}
|
|
|
|
// openEnvelope tries to decrypt an envelope, first based on asymetric key (if
|
|
// provided) and second on the symetric key (if provided)
|
|
func (s *WakuMailServer) openEnvelope(request *wakucommon.Envelope) *wakucommon.ReceivedMessage {
|
|
if s.asymFilter != nil {
|
|
if d := request.Open(s.asymFilter); d != nil {
|
|
return d
|
|
}
|
|
}
|
|
if s.symFilter != nil {
|
|
if d := request.Open(s.symFilter); d != nil {
|
|
return d
|
|
}
|
|
}
|
|
return nil
|
|
}
|
|
|
|
func (s *WakuMailServer) decodeRequest(peerID []byte, request *wakucommon.Envelope) (MessagesRequestPayload, error) {
|
|
var payload MessagesRequestPayload
|
|
|
|
if s.minRequestPoW > 0.0 && request.PoW() < s.minRequestPoW {
|
|
return payload, errors.New("PoW too low")
|
|
}
|
|
|
|
decrypted := s.openEnvelope(request)
|
|
if decrypted == nil {
|
|
logutils.ZapLogger().Warn("Failed to decrypt p2p request")
|
|
return payload, errors.New("failed to decrypt p2p request")
|
|
}
|
|
|
|
if err := checkMsgSignature(decrypted.Src, peerID); err != nil {
|
|
logutils.ZapLogger().Warn("Check message signature failed", zap.Error(err))
|
|
return payload, fmt.Errorf("check message signature failed: %v", err)
|
|
}
|
|
|
|
if err := rlp.DecodeBytes(decrypted.Payload, &payload); err != nil {
|
|
return payload, fmt.Errorf("failed to decode data: %v", err)
|
|
}
|
|
|
|
if payload.Upper == 0 {
|
|
payload.Upper = uint32(time.Now().Unix() + whisperTTLSafeThreshold)
|
|
}
|
|
|
|
if payload.Upper < payload.Lower {
|
|
logutils.ZapLogger().Error("Query range is invalid: lower > upper", zap.Uint32("lower", payload.Lower), zap.Uint32("upper", payload.Upper))
|
|
return payload, errors.New("query range is invalid: lower > upper")
|
|
}
|
|
|
|
return payload, nil
|
|
}
|
|
|
|
// -------
|
|
// adapter
|
|
// -------
|
|
|
|
type adapter interface {
|
|
CreateRequestFailedPayload(reqID types.Hash, err error) []byte
|
|
CreateRequestCompletedPayload(reqID, lastEnvelopeHash types.Hash, cursor []byte) []byte
|
|
CreateSyncResponse(envelopes []types.Envelope, cursor []byte, final bool, err string) interface{}
|
|
CreateRawSyncResponse(envelopes []rlp.RawValue, cursor []byte, final bool, err string) interface{}
|
|
}
|
|
|
|
// -----------
|
|
// wakuAdapter
|
|
// -----------
|
|
|
|
type wakuAdapter struct{}
|
|
|
|
var _ adapter = (*wakuAdapter)(nil)
|
|
|
|
func (wakuAdapter) CreateRequestFailedPayload(reqID types.Hash, err error) []byte {
|
|
return waku.CreateMailServerRequestFailedPayload(common.Hash(reqID), err)
|
|
}
|
|
|
|
func (wakuAdapter) CreateRequestCompletedPayload(reqID, lastEnvelopeHash types.Hash, cursor []byte) []byte {
|
|
return waku.CreateMailServerRequestCompletedPayload(common.Hash(reqID), common.Hash(lastEnvelopeHash), cursor)
|
|
}
|
|
|
|
func (wakuAdapter) CreateSyncResponse(_ []types.Envelope, _ []byte, _ bool, _ string) interface{} {
|
|
return nil
|
|
}
|
|
|
|
func (wakuAdapter) CreateRawSyncResponse(_ []rlp.RawValue, _ []byte, _ bool, _ string) interface{} {
|
|
return nil
|
|
}
|
|
|
|
// -------
|
|
// service
|
|
// -------
|
|
|
|
type service interface {
|
|
SendHistoricMessageResponse(peerID []byte, payload []byte) error
|
|
SendRawP2PDirect(peerID []byte, envelopes ...rlp.RawValue) error
|
|
MaxMessageSize() uint32
|
|
SendRawSyncResponse(peerID []byte, data interface{}) error // optional
|
|
SendSyncResponse(peerID []byte, data interface{}) error // optional
|
|
}
|
|
|
|
// -----------
|
|
// wakuService
|
|
// -----------
|
|
|
|
type wakuService struct {
|
|
*waku.Waku
|
|
}
|
|
|
|
func (s *wakuService) SendRawSyncResponse(peerID []byte, data interface{}) error {
|
|
return errors.New("syncing mailservers is not support by Waku")
|
|
}
|
|
|
|
func (s *wakuService) SendSyncResponse(peerID []byte, data interface{}) error {
|
|
return errors.New("syncing mailservers is not support by Waku")
|
|
}
|
|
|
|
// ----------
|
|
// mailServer
|
|
// ----------
|
|
|
|
type mailServer struct {
|
|
adapter adapter
|
|
service service
|
|
db DB
|
|
cleaner *dbCleaner // removes old envelopes
|
|
muRateLimiter sync.RWMutex
|
|
rateLimiter *rateLimiter
|
|
}
|
|
|
|
func newMailServer(cfg Config, adapter adapter, service service) (*mailServer, error) {
|
|
if len(cfg.DataDir) == 0 {
|
|
return nil, errDirectoryNotProvided
|
|
}
|
|
|
|
// TODO: move out
|
|
if len(cfg.Password) == 0 && len(cfg.AsymKey) == 0 {
|
|
return nil, errDecryptionMethodNotProvided
|
|
}
|
|
|
|
s := mailServer{
|
|
adapter: adapter,
|
|
service: service,
|
|
}
|
|
|
|
if cfg.RateLimit > 0 {
|
|
s.setupRateLimiter(time.Duration(cfg.RateLimit) * time.Second)
|
|
}
|
|
|
|
// Open database in the last step in order not to init with error
|
|
// and leave the database open by accident.
|
|
if cfg.PostgresEnabled {
|
|
logutils.ZapLogger().Info("Connecting to postgres database")
|
|
database, err := NewPostgresDB(cfg.PostgresURI)
|
|
if err != nil {
|
|
return nil, fmt.Errorf("open DB: %s", err)
|
|
}
|
|
s.db = database
|
|
logutils.ZapLogger().Info("Connected to postgres database")
|
|
} else {
|
|
// Defaults to LevelDB
|
|
database, err := NewLevelDB(cfg.DataDir)
|
|
if err != nil {
|
|
return nil, fmt.Errorf("open DB: %s", err)
|
|
}
|
|
s.db = database
|
|
}
|
|
|
|
if cfg.DataRetention > 0 {
|
|
// MailServerDataRetention is a number of days.
|
|
s.setupCleaner(time.Duration(cfg.DataRetention) * time.Hour * 24)
|
|
}
|
|
|
|
return &s, nil
|
|
}
|
|
|
|
// setupRateLimiter in case limit is bigger than 0 it will setup an automated
|
|
// limit db cleanup.
|
|
func (s *mailServer) setupRateLimiter(limit time.Duration) {
|
|
s.rateLimiter = newRateLimiter(limit)
|
|
s.rateLimiter.Start()
|
|
}
|
|
|
|
func (s *mailServer) setupCleaner(retention time.Duration) {
|
|
s.cleaner = newDBCleaner(s.db, retention)
|
|
s.cleaner.Start()
|
|
}
|
|
|
|
func (s *mailServer) Archive(env types.Envelope) {
|
|
err := s.db.SaveEnvelope(env)
|
|
if err != nil {
|
|
logutils.ZapLogger().Error("Could not save envelope", zap.Stringer("hash", env.Hash()))
|
|
}
|
|
}
|
|
|
|
func (s *mailServer) DeliverMail(peerID, reqID types.Hash, req MessagesRequestPayload) {
|
|
timer := prom.NewTimer(mailDeliveryDuration)
|
|
defer timer.ObserveDuration()
|
|
|
|
deliveryAttemptsCounter.Inc()
|
|
logutils.ZapLogger().Info(
|
|
"[mailserver:DeliverMail] delivering mail",
|
|
zap.Stringer("peerID", peerID),
|
|
zap.Stringer("requestID", reqID),
|
|
)
|
|
|
|
req.SetDefaults()
|
|
|
|
logutils.ZapLogger().Info(
|
|
"[mailserver:DeliverMail] processing request",
|
|
zap.Stringer("peerID", peerID),
|
|
zap.Stringer("requestID", reqID),
|
|
zap.Uint32("lower", req.Lower),
|
|
zap.Uint32("upper", req.Upper),
|
|
zap.Binary("bloom", req.Bloom),
|
|
zap.Any("topics", req.Topics),
|
|
zap.Uint32("limit", req.Limit),
|
|
zap.Binary("cursor", req.Cursor),
|
|
zap.Bool("batch", req.Batch),
|
|
)
|
|
|
|
if err := req.Validate(); err != nil {
|
|
syncFailuresCounter.WithLabelValues("req_invalid").Inc()
|
|
logutils.ZapLogger().Error(
|
|
"[mailserver:DeliverMail] request invalid",
|
|
zap.Stringer("peerID", peerID),
|
|
zap.Stringer("requestID", reqID),
|
|
zap.Error(err),
|
|
)
|
|
s.sendHistoricMessageErrorResponse(peerID, reqID, fmt.Errorf("request is invalid: %v", err))
|
|
return
|
|
}
|
|
|
|
if s.exceedsPeerRequests(peerID) {
|
|
deliveryFailuresCounter.WithLabelValues("peer_req_limit").Inc()
|
|
logutils.ZapLogger().Error(
|
|
"[mailserver:DeliverMail] peer exceeded the limit",
|
|
zap.Stringer("peerID", peerID),
|
|
zap.Stringer("requestID", reqID),
|
|
)
|
|
s.sendHistoricMessageErrorResponse(peerID, reqID, fmt.Errorf("rate limit exceeded"))
|
|
return
|
|
}
|
|
|
|
if req.Batch {
|
|
requestsBatchedCounter.Inc()
|
|
}
|
|
|
|
iter, err := s.createIterator(req)
|
|
if err != nil {
|
|
logutils.ZapLogger().Error(
|
|
"[mailserver:DeliverMail] request failed",
|
|
zap.Stringer("peerID", peerID),
|
|
zap.Stringer("requestID", reqID),
|
|
zap.Error(err),
|
|
)
|
|
return
|
|
}
|
|
defer func() { _ = iter.Release() }()
|
|
|
|
bundles := make(chan []rlp.RawValue, 5)
|
|
errCh := make(chan error)
|
|
cancelProcessing := make(chan struct{})
|
|
|
|
go func() {
|
|
defer gocommon.LogOnPanic()
|
|
counter := 0
|
|
for bundle := range bundles {
|
|
if err := s.sendRawEnvelopes(peerID, bundle, req.Batch); err != nil {
|
|
close(cancelProcessing)
|
|
errCh <- err
|
|
break
|
|
}
|
|
counter++
|
|
}
|
|
close(errCh)
|
|
logutils.ZapLogger().Info(
|
|
"[mailserver:DeliverMail] finished sending bundles",
|
|
zap.Stringer("peerID", peerID),
|
|
zap.Stringer("requestID", reqID),
|
|
zap.Int("counter", counter),
|
|
)
|
|
}()
|
|
|
|
nextPageCursor, lastEnvelopeHash := s.processRequestInBundles(
|
|
iter,
|
|
req.Bloom,
|
|
req.Topics,
|
|
int(req.Limit),
|
|
processRequestTimeout,
|
|
reqID.String(),
|
|
bundles,
|
|
cancelProcessing,
|
|
)
|
|
|
|
// Wait for the goroutine to finish the work. It may return an error.
|
|
if err := <-errCh; err != nil {
|
|
deliveryFailuresCounter.WithLabelValues("process").Inc()
|
|
logutils.ZapLogger().Error(
|
|
"[mailserver:DeliverMail] error while processing",
|
|
zap.Stringer("peerID", peerID),
|
|
zap.Stringer("requestID", reqID),
|
|
zap.Error(err),
|
|
)
|
|
s.sendHistoricMessageErrorResponse(peerID, reqID, err)
|
|
return
|
|
}
|
|
|
|
// Processing of the request could be finished earlier due to iterator error.
|
|
if err := iter.Error(); err != nil {
|
|
deliveryFailuresCounter.WithLabelValues("iterator").Inc()
|
|
logutils.ZapLogger().Error(
|
|
"[mailserver:DeliverMail] iterator failed",
|
|
zap.Stringer("peerID", peerID),
|
|
zap.Stringer("requestID", reqID),
|
|
zap.Error(err),
|
|
)
|
|
s.sendHistoricMessageErrorResponse(peerID, reqID, err)
|
|
return
|
|
}
|
|
|
|
logutils.ZapLogger().Info(
|
|
"[mailserver:DeliverMail] sending historic message response",
|
|
zap.Stringer("peerID", peerID),
|
|
zap.Stringer("requestID", reqID),
|
|
zap.Stringer("last", lastEnvelopeHash),
|
|
zap.Binary("next", nextPageCursor),
|
|
)
|
|
|
|
s.sendHistoricMessageResponse(peerID, reqID, lastEnvelopeHash, nextPageCursor)
|
|
}
|
|
|
|
func (s *mailServer) SyncMail(peerID types.Hash, req MessagesRequestPayload) error {
|
|
logutils.ZapLogger().Info("Started syncing envelopes", zap.Stringer("peer", peerID), zap.Any("req", req))
|
|
|
|
requestID := fmt.Sprintf("%d-%d", time.Now().UnixNano(), rand.Intn(1000)) // nolint: gosec
|
|
|
|
syncAttemptsCounter.Inc()
|
|
|
|
// Check rate limiting for a requesting peer.
|
|
if s.exceedsPeerRequests(peerID) {
|
|
syncFailuresCounter.WithLabelValues("req_per_sec_limit").Inc()
|
|
logutils.ZapLogger().Error("Peer exceeded request per seconds limit", zap.Stringer("peerID", peerID))
|
|
return fmt.Errorf("requests per seconds limit exceeded")
|
|
}
|
|
|
|
req.SetDefaults()
|
|
|
|
if err := req.Validate(); err != nil {
|
|
syncFailuresCounter.WithLabelValues("req_invalid").Inc()
|
|
return fmt.Errorf("request is invalid: %v", err)
|
|
}
|
|
|
|
iter, err := s.createIterator(req)
|
|
if err != nil {
|
|
syncFailuresCounter.WithLabelValues("iterator").Inc()
|
|
return err
|
|
}
|
|
defer func() { _ = iter.Release() }()
|
|
|
|
bundles := make(chan []rlp.RawValue, 5)
|
|
errCh := make(chan error)
|
|
cancelProcessing := make(chan struct{})
|
|
|
|
go func() {
|
|
defer gocommon.LogOnPanic()
|
|
for bundle := range bundles {
|
|
resp := s.adapter.CreateRawSyncResponse(bundle, nil, false, "")
|
|
if err := s.service.SendRawSyncResponse(peerID.Bytes(), resp); err != nil {
|
|
close(cancelProcessing)
|
|
errCh <- fmt.Errorf("failed to send sync response: %v", err)
|
|
break
|
|
}
|
|
}
|
|
close(errCh)
|
|
}()
|
|
|
|
nextCursor, _ := s.processRequestInBundles(
|
|
iter,
|
|
req.Bloom,
|
|
req.Topics,
|
|
int(req.Limit),
|
|
processRequestTimeout,
|
|
requestID,
|
|
bundles,
|
|
cancelProcessing,
|
|
)
|
|
|
|
// Wait for the goroutine to finish the work. It may return an error.
|
|
if err := <-errCh; err != nil {
|
|
syncFailuresCounter.WithLabelValues("routine").Inc()
|
|
_ = s.service.SendSyncResponse(
|
|
peerID.Bytes(),
|
|
s.adapter.CreateSyncResponse(nil, nil, false, "failed to send a response"),
|
|
)
|
|
return err
|
|
}
|
|
|
|
// Processing of the request could be finished earlier due to iterator error.
|
|
if err := iter.Error(); err != nil {
|
|
syncFailuresCounter.WithLabelValues("iterator").Inc()
|
|
_ = s.service.SendSyncResponse(
|
|
peerID.Bytes(),
|
|
s.adapter.CreateSyncResponse(nil, nil, false, "failed to process all envelopes"),
|
|
)
|
|
return fmt.Errorf("LevelDB iterator failed: %v", err)
|
|
}
|
|
|
|
logutils.ZapLogger().Info("Finished syncing envelopes", zap.Stringer("peer", peerID))
|
|
|
|
err = s.service.SendSyncResponse(
|
|
peerID.Bytes(),
|
|
s.adapter.CreateSyncResponse(nil, nextCursor, true, ""),
|
|
)
|
|
if err != nil {
|
|
syncFailuresCounter.WithLabelValues("response_send").Inc()
|
|
return fmt.Errorf("failed to send the final sync response: %v", err)
|
|
}
|
|
|
|
return nil
|
|
}
|
|
|
|
// Close the mailserver and its associated db connection.
|
|
func (s *mailServer) Close() {
|
|
if s.db != nil {
|
|
if err := s.db.Close(); err != nil {
|
|
logutils.ZapLogger().Error("closing database failed", zap.Error(err))
|
|
}
|
|
}
|
|
if s.rateLimiter != nil {
|
|
s.rateLimiter.Stop()
|
|
}
|
|
if s.cleaner != nil {
|
|
s.cleaner.Stop()
|
|
}
|
|
}
|
|
|
|
func (s *mailServer) exceedsPeerRequests(peerID types.Hash) bool {
|
|
s.muRateLimiter.RLock()
|
|
defer s.muRateLimiter.RUnlock()
|
|
|
|
if s.rateLimiter == nil {
|
|
return false
|
|
}
|
|
|
|
if s.rateLimiter.IsAllowed(peerID.String()) {
|
|
s.rateLimiter.Add(peerID.String())
|
|
return false
|
|
}
|
|
|
|
logutils.ZapLogger().Info("peerID exceeded the number of requests per second", zap.Stringer("peerID", peerID))
|
|
return true
|
|
}
|
|
|
|
func (s *mailServer) createIterator(req MessagesRequestPayload) (Iterator, error) {
|
|
var (
|
|
emptyHash types.Hash
|
|
emptyTopic types.TopicType
|
|
ku, kl *DBKey
|
|
)
|
|
|
|
ku = NewDBKey(req.Upper+1, emptyTopic, emptyHash)
|
|
kl = NewDBKey(req.Lower, emptyTopic, emptyHash)
|
|
|
|
query := CursorQuery{
|
|
start: kl.Bytes(),
|
|
end: ku.Bytes(),
|
|
cursor: req.Cursor,
|
|
topics: req.Topics,
|
|
bloom: req.Bloom,
|
|
limit: req.Limit,
|
|
}
|
|
return s.db.BuildIterator(query)
|
|
}
|
|
|
|
func (s *mailServer) processRequestInBundles(
|
|
iter Iterator,
|
|
bloom []byte,
|
|
topics [][]byte,
|
|
limit int,
|
|
timeout time.Duration,
|
|
requestID string,
|
|
output chan<- []rlp.RawValue,
|
|
cancel <-chan struct{},
|
|
) ([]byte, types.Hash) {
|
|
timer := prom.NewTimer(requestsInBundlesDuration)
|
|
defer timer.ObserveDuration()
|
|
|
|
var (
|
|
bundle []rlp.RawValue
|
|
bundleSize uint32
|
|
batches [][]rlp.RawValue
|
|
processedEnvelopes int
|
|
processedEnvelopesSize int64
|
|
nextCursor []byte
|
|
lastEnvelopeHash types.Hash
|
|
)
|
|
|
|
logutils.ZapLogger().Info(
|
|
"[mailserver:processRequestInBundles] processing request",
|
|
zap.String("requestID", requestID),
|
|
zap.Int("limit", limit),
|
|
)
|
|
|
|
var topicsMap map[types.TopicType]bool
|
|
|
|
if len(topics) != 0 {
|
|
topicsMap = make(map[types.TopicType]bool)
|
|
for _, t := range topics {
|
|
topicsMap[types.BytesToTopic(t)] = true
|
|
}
|
|
}
|
|
|
|
// We iterate over the envelopes.
|
|
// We collect envelopes in batches.
|
|
// If there still room and we haven't reached the limit
|
|
// append and continue.
|
|
// Otherwise publish what you have so far, reset the bundle to the
|
|
// current envelope, and leave if we hit the limit
|
|
for iter.Next() {
|
|
var rawValue []byte
|
|
var err error
|
|
if len(topicsMap) != 0 {
|
|
rawValue, err = iter.GetEnvelopeByTopicsMap(topicsMap)
|
|
|
|
} else if len(bloom) != 0 {
|
|
rawValue, err = iter.GetEnvelopeByBloomFilter(bloom)
|
|
} else {
|
|
err = errors.New("either topics or bloom must be specified")
|
|
}
|
|
if err != nil {
|
|
logutils.ZapLogger().Error(
|
|
"[mailserver:processRequestInBundles]Failed to get envelope from iterator",
|
|
zap.String("requestID", requestID),
|
|
zap.Error(err),
|
|
)
|
|
continue
|
|
}
|
|
|
|
if rawValue == nil {
|
|
continue
|
|
}
|
|
|
|
key, err := iter.DBKey()
|
|
if err != nil {
|
|
logutils.ZapLogger().Error(
|
|
"[mailserver:processRequestInBundles] failed getting key",
|
|
zap.String("requestID", requestID),
|
|
zap.Error(err),
|
|
)
|
|
break
|
|
|
|
}
|
|
|
|
// TODO(adam): this is invalid code. If the limit is 1000,
|
|
// it will only send 999 items and send a cursor.
|
|
lastEnvelopeHash = key.EnvelopeHash()
|
|
processedEnvelopes++
|
|
envelopeSize := uint32(len(rawValue))
|
|
limitReached := processedEnvelopes >= limit
|
|
newSize := bundleSize + envelopeSize
|
|
|
|
// If we still have some room for messages, add and continue
|
|
if !limitReached && newSize < s.service.MaxMessageSize() {
|
|
bundle = append(bundle, rawValue)
|
|
bundleSize = newSize
|
|
continue
|
|
}
|
|
|
|
// Publish if anything is in the bundle (there should always be
|
|
// something unless limit = 1)
|
|
if len(bundle) != 0 {
|
|
batches = append(batches, bundle)
|
|
processedEnvelopesSize += int64(bundleSize)
|
|
}
|
|
|
|
// Reset the bundle with the current envelope
|
|
bundle = []rlp.RawValue{rawValue}
|
|
bundleSize = envelopeSize
|
|
|
|
// Leave if we reached the limit
|
|
if limitReached {
|
|
nextCursor = key.Cursor()
|
|
break
|
|
}
|
|
}
|
|
|
|
if len(bundle) > 0 {
|
|
batches = append(batches, bundle)
|
|
processedEnvelopesSize += int64(bundleSize)
|
|
}
|
|
|
|
logutils.ZapLogger().Info(
|
|
"[mailserver:processRequestInBundles] publishing envelopes",
|
|
zap.String("requestID", requestID),
|
|
zap.Int("batchesCount", len(batches)),
|
|
zap.Int("envelopeCount", processedEnvelopes),
|
|
zap.Int64("processedEnvelopesSize", processedEnvelopesSize),
|
|
zap.Binary("cursor", nextCursor),
|
|
)
|
|
|
|
// Publish
|
|
batchLoop:
|
|
for _, batch := range batches {
|
|
select {
|
|
case output <- batch:
|
|
// It might happen that during producing the batches,
|
|
// the connection with the peer goes down and
|
|
// the consumer of `output` channel exits prematurely.
|
|
// In such a case, we should stop pushing batches and exit.
|
|
case <-cancel:
|
|
logutils.ZapLogger().Info(
|
|
"[mailserver:processRequestInBundles] failed to push all batches",
|
|
zap.String("requestID", requestID),
|
|
)
|
|
break batchLoop
|
|
case <-time.After(timeout):
|
|
logutils.ZapLogger().Error(
|
|
"[mailserver:processRequestInBundles] timed out pushing a batch",
|
|
zap.String("requestID", requestID),
|
|
)
|
|
break batchLoop
|
|
}
|
|
}
|
|
|
|
envelopesCounter.Inc()
|
|
sentEnvelopeBatchSizeMeter.Observe(float64(processedEnvelopesSize))
|
|
|
|
logutils.ZapLogger().Info(
|
|
"[mailserver:processRequestInBundles] envelopes published",
|
|
zap.String("requestID", requestID),
|
|
)
|
|
close(output)
|
|
|
|
return nextCursor, lastEnvelopeHash
|
|
}
|
|
|
|
func (s *mailServer) sendRawEnvelopes(peerID types.Hash, envelopes []rlp.RawValue, batch bool) error {
|
|
timer := prom.NewTimer(sendRawEnvelopeDuration)
|
|
defer timer.ObserveDuration()
|
|
|
|
if batch {
|
|
return s.service.SendRawP2PDirect(peerID.Bytes(), envelopes...)
|
|
}
|
|
|
|
for _, env := range envelopes {
|
|
if err := s.service.SendRawP2PDirect(peerID.Bytes(), env); err != nil {
|
|
return err
|
|
}
|
|
}
|
|
|
|
return nil
|
|
}
|
|
|
|
func (s *mailServer) sendHistoricMessageResponse(peerID, reqID, lastEnvelopeHash types.Hash, cursor []byte) {
|
|
payload := s.adapter.CreateRequestCompletedPayload(reqID, lastEnvelopeHash, cursor)
|
|
err := s.service.SendHistoricMessageResponse(peerID.Bytes(), payload)
|
|
if err != nil {
|
|
deliveryFailuresCounter.WithLabelValues("historic_msg_resp").Inc()
|
|
logutils.ZapLogger().Error(
|
|
"[mailserver:DeliverMail] error sending historic message response",
|
|
zap.Stringer("peerID", peerID),
|
|
zap.Stringer("requestID", reqID),
|
|
zap.Error(err),
|
|
)
|
|
}
|
|
}
|
|
|
|
func (s *mailServer) sendHistoricMessageErrorResponse(peerID, reqID types.Hash, errorToReport error) {
|
|
payload := s.adapter.CreateRequestFailedPayload(reqID, errorToReport)
|
|
err := s.service.SendHistoricMessageResponse(peerID.Bytes(), payload)
|
|
// if we can't report an error, probably something is wrong with p2p connection,
|
|
// so we just print a log entry to document this sad fact
|
|
if err != nil {
|
|
logutils.ZapLogger().Error("Error while reporting error response", zap.Stringer("peerID", peerID), zap.Error(err))
|
|
}
|
|
}
|
|
|
|
func extractBloomFromEncodedEnvelope(rawValue rlp.RawValue) ([]byte, error) {
|
|
var envelope wakucommon.Envelope
|
|
decodeErr := rlp.DecodeBytes(rawValue, &envelope)
|
|
if decodeErr != nil {
|
|
return nil, decodeErr
|
|
}
|
|
return envelope.Bloom(), nil
|
|
}
|
|
|
|
// checkMsgSignature returns an error in case the message is not correctly signed.
|
|
func checkMsgSignature(reqSrc *ecdsa.PublicKey, id []byte) error {
|
|
src := crypto.FromECDSAPub(reqSrc)
|
|
if len(src)-len(id) == 1 {
|
|
src = src[1:]
|
|
}
|
|
|
|
// if you want to check the signature, you can do it here. e.g.:
|
|
// if !bytes.Equal(peerID, src) {
|
|
if src == nil {
|
|
return errors.New("wrong signature of p2p request")
|
|
}
|
|
|
|
return nil
|
|
}
|