mirror of https://github.com/status-im/op-geth.git
665 lines
19 KiB
Go
665 lines
19 KiB
Go
// Copyright 2015 The go-ethereum Authors
|
|
// This file is part of the go-ethereum library.
|
|
//
|
|
// The go-ethereum library is free software: you can redistribute it and/or modify
|
|
// it under the terms of the GNU Lesser General Public License as published by
|
|
// the Free Software Foundation, either version 3 of the License, or
|
|
// (at your option) any later version.
|
|
//
|
|
// The go-ethereum library is distributed in the hope that it will be useful,
|
|
// but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
// GNU Lesser General Public License for more details.
|
|
//
|
|
// You should have received a copy of the GNU Lesser General Public License
|
|
// along with the go-ethereum library. If not, see <http://www.gnu.org/licenses/>.
|
|
|
|
package node
|
|
|
|
import (
|
|
"errors"
|
|
"fmt"
|
|
"net"
|
|
"os"
|
|
"path/filepath"
|
|
"reflect"
|
|
"strings"
|
|
"sync"
|
|
|
|
"github.com/ethereum/go-ethereum/accounts"
|
|
"github.com/ethereum/go-ethereum/core/rawdb"
|
|
"github.com/ethereum/go-ethereum/ethdb"
|
|
"github.com/ethereum/go-ethereum/event"
|
|
"github.com/ethereum/go-ethereum/internal/debug"
|
|
"github.com/ethereum/go-ethereum/log"
|
|
"github.com/ethereum/go-ethereum/p2p"
|
|
"github.com/ethereum/go-ethereum/rpc"
|
|
"github.com/prometheus/tsdb/fileutil"
|
|
)
|
|
|
|
// Node is a container on which services can be registered.
|
|
type Node struct {
|
|
eventmux *event.TypeMux // Event multiplexer used between the services of a stack
|
|
config *Config
|
|
accman *accounts.Manager
|
|
|
|
ephemeralKeystore string // if non-empty, the key directory that will be removed by Stop
|
|
instanceDirLock fileutil.Releaser // prevents concurrent use of instance directory
|
|
|
|
serverConfig p2p.Config
|
|
server *p2p.Server // Currently running P2P networking layer
|
|
|
|
serviceFuncs []ServiceConstructor // Service constructors (in dependency order)
|
|
services map[reflect.Type]Service // Currently running services
|
|
|
|
rpcAPIs []rpc.API // List of APIs currently provided by the node
|
|
inprocHandler *rpc.Server // In-process RPC request handler to process the API requests
|
|
|
|
ipcEndpoint string // IPC endpoint to listen at (empty = IPC disabled)
|
|
ipcListener net.Listener // IPC RPC listener socket to serve API requests
|
|
ipcHandler *rpc.Server // IPC RPC request handler to process the API requests
|
|
|
|
httpEndpoint string // HTTP endpoint (interface + port) to listen at (empty = HTTP disabled)
|
|
httpWhitelist []string // HTTP RPC modules to allow through this endpoint
|
|
httpListener net.Listener // HTTP RPC listener socket to server API requests
|
|
httpHandler *rpc.Server // HTTP RPC request handler to process the API requests
|
|
|
|
wsEndpoint string // Websocket endpoint (interface + port) to listen at (empty = websocket disabled)
|
|
wsListener net.Listener // Websocket RPC listener socket to server API requests
|
|
wsHandler *rpc.Server // Websocket RPC request handler to process the API requests
|
|
|
|
stop chan struct{} // Channel to wait for termination notifications
|
|
lock sync.RWMutex
|
|
|
|
log log.Logger
|
|
}
|
|
|
|
// New creates a new P2P node, ready for protocol registration.
|
|
func New(conf *Config) (*Node, error) {
|
|
// Copy config and resolve the datadir so future changes to the current
|
|
// working directory don't affect the node.
|
|
confCopy := *conf
|
|
conf = &confCopy
|
|
if conf.DataDir != "" {
|
|
absdatadir, err := filepath.Abs(conf.DataDir)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
conf.DataDir = absdatadir
|
|
}
|
|
// Ensure that the instance name doesn't cause weird conflicts with
|
|
// other files in the data directory.
|
|
if strings.ContainsAny(conf.Name, `/\`) {
|
|
return nil, errors.New(`Config.Name must not contain '/' or '\'`)
|
|
}
|
|
if conf.Name == datadirDefaultKeyStore {
|
|
return nil, errors.New(`Config.Name cannot be "` + datadirDefaultKeyStore + `"`)
|
|
}
|
|
if strings.HasSuffix(conf.Name, ".ipc") {
|
|
return nil, errors.New(`Config.Name cannot end in ".ipc"`)
|
|
}
|
|
// Ensure that the AccountManager method works before the node has started.
|
|
// We rely on this in cmd/geth.
|
|
am, ephemeralKeystore, err := makeAccountManager(conf)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
if conf.Logger == nil {
|
|
conf.Logger = log.New()
|
|
}
|
|
// Note: any interaction with Config that would create/touch files
|
|
// in the data directory or instance directory is delayed until Start.
|
|
return &Node{
|
|
accman: am,
|
|
ephemeralKeystore: ephemeralKeystore,
|
|
config: conf,
|
|
serviceFuncs: []ServiceConstructor{},
|
|
ipcEndpoint: conf.IPCEndpoint(),
|
|
httpEndpoint: conf.HTTPEndpoint(),
|
|
wsEndpoint: conf.WSEndpoint(),
|
|
eventmux: new(event.TypeMux),
|
|
log: conf.Logger,
|
|
}, nil
|
|
}
|
|
|
|
// Close stops the Node and releases resources acquired in
|
|
// Node constructor New.
|
|
func (n *Node) Close() error {
|
|
var errs []error
|
|
|
|
// Terminate all subsystems and collect any errors
|
|
if err := n.Stop(); err != nil && err != ErrNodeStopped {
|
|
errs = append(errs, err)
|
|
}
|
|
if err := n.accman.Close(); err != nil {
|
|
errs = append(errs, err)
|
|
}
|
|
// Report any errors that might have occurred
|
|
switch len(errs) {
|
|
case 0:
|
|
return nil
|
|
case 1:
|
|
return errs[0]
|
|
default:
|
|
return fmt.Errorf("%v", errs)
|
|
}
|
|
}
|
|
|
|
// Register injects a new service into the node's stack. The service created by
|
|
// the passed constructor must be unique in its type with regard to sibling ones.
|
|
func (n *Node) Register(constructor ServiceConstructor) error {
|
|
n.lock.Lock()
|
|
defer n.lock.Unlock()
|
|
|
|
if n.server != nil {
|
|
return ErrNodeRunning
|
|
}
|
|
n.serviceFuncs = append(n.serviceFuncs, constructor)
|
|
return nil
|
|
}
|
|
|
|
// Start create a live P2P node and starts running it.
|
|
func (n *Node) Start() error {
|
|
n.lock.Lock()
|
|
defer n.lock.Unlock()
|
|
|
|
// Short circuit if the node's already running
|
|
if n.server != nil {
|
|
return ErrNodeRunning
|
|
}
|
|
if err := n.openDataDir(); err != nil {
|
|
return err
|
|
}
|
|
|
|
// Initialize the p2p server. This creates the node key and
|
|
// discovery databases.
|
|
n.serverConfig = n.config.P2P
|
|
n.serverConfig.PrivateKey = n.config.NodeKey()
|
|
n.serverConfig.Name = n.config.NodeName()
|
|
n.serverConfig.Logger = n.log
|
|
if n.serverConfig.StaticNodes == nil {
|
|
n.serverConfig.StaticNodes = n.config.StaticNodes()
|
|
}
|
|
if n.serverConfig.TrustedNodes == nil {
|
|
n.serverConfig.TrustedNodes = n.config.TrustedNodes()
|
|
}
|
|
if n.serverConfig.NodeDatabase == "" {
|
|
n.serverConfig.NodeDatabase = n.config.NodeDB()
|
|
}
|
|
running := &p2p.Server{Config: n.serverConfig}
|
|
n.log.Info("Starting peer-to-peer node", "instance", n.serverConfig.Name)
|
|
|
|
// Otherwise copy and specialize the P2P configuration
|
|
services := make(map[reflect.Type]Service)
|
|
for _, constructor := range n.serviceFuncs {
|
|
// Create a new context for the particular service
|
|
ctx := &ServiceContext{
|
|
config: n.config,
|
|
services: make(map[reflect.Type]Service),
|
|
EventMux: n.eventmux,
|
|
AccountManager: n.accman,
|
|
}
|
|
for kind, s := range services { // copy needed for threaded access
|
|
ctx.services[kind] = s
|
|
}
|
|
// Construct and save the service
|
|
service, err := constructor(ctx)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
kind := reflect.TypeOf(service)
|
|
if _, exists := services[kind]; exists {
|
|
return &DuplicateServiceError{Kind: kind}
|
|
}
|
|
services[kind] = service
|
|
}
|
|
// Gather the protocols and start the freshly assembled P2P server
|
|
for _, service := range services {
|
|
running.Protocols = append(running.Protocols, service.Protocols()...)
|
|
}
|
|
if err := running.Start(); err != nil {
|
|
return convertFileLockError(err)
|
|
}
|
|
// Start each of the services
|
|
var started []reflect.Type
|
|
for kind, service := range services {
|
|
// Start the next service, stopping all previous upon failure
|
|
if err := service.Start(running); err != nil {
|
|
for _, kind := range started {
|
|
services[kind].Stop()
|
|
}
|
|
running.Stop()
|
|
|
|
return err
|
|
}
|
|
// Mark the service started for potential cleanup
|
|
started = append(started, kind)
|
|
}
|
|
// Lastly start the configured RPC interfaces
|
|
if err := n.startRPC(services); err != nil {
|
|
for _, service := range services {
|
|
service.Stop()
|
|
}
|
|
running.Stop()
|
|
return err
|
|
}
|
|
// Finish initializing the startup
|
|
n.services = services
|
|
n.server = running
|
|
n.stop = make(chan struct{})
|
|
return nil
|
|
}
|
|
|
|
// Config returns the configuration of node.
|
|
func (n *Node) Config() *Config {
|
|
return n.config
|
|
}
|
|
|
|
func (n *Node) openDataDir() error {
|
|
if n.config.DataDir == "" {
|
|
return nil // ephemeral
|
|
}
|
|
|
|
instdir := filepath.Join(n.config.DataDir, n.config.name())
|
|
if err := os.MkdirAll(instdir, 0700); err != nil {
|
|
return err
|
|
}
|
|
// Lock the instance directory to prevent concurrent use by another instance as well as
|
|
// accidental use of the instance directory as a database.
|
|
release, _, err := fileutil.Flock(filepath.Join(instdir, "LOCK"))
|
|
if err != nil {
|
|
return convertFileLockError(err)
|
|
}
|
|
n.instanceDirLock = release
|
|
return nil
|
|
}
|
|
|
|
// startRPC is a helper method to start all the various RPC endpoint during node
|
|
// startup. It's not meant to be called at any time afterwards as it makes certain
|
|
// assumptions about the state of the node.
|
|
func (n *Node) startRPC(services map[reflect.Type]Service) error {
|
|
// Gather all the possible APIs to surface
|
|
apis := n.apis()
|
|
for _, service := range services {
|
|
apis = append(apis, service.APIs()...)
|
|
}
|
|
// Start the various API endpoints, terminating all in case of errors
|
|
if err := n.startInProc(apis); err != nil {
|
|
return err
|
|
}
|
|
if err := n.startIPC(apis); err != nil {
|
|
n.stopInProc()
|
|
return err
|
|
}
|
|
if err := n.startHTTP(n.httpEndpoint, apis, n.config.HTTPModules, n.config.HTTPCors, n.config.HTTPVirtualHosts, n.config.HTTPTimeouts); err != nil {
|
|
n.stopIPC()
|
|
n.stopInProc()
|
|
return err
|
|
}
|
|
if err := n.startWS(n.wsEndpoint, apis, n.config.WSModules, n.config.WSOrigins, n.config.WSExposeAll); err != nil {
|
|
n.stopHTTP()
|
|
n.stopIPC()
|
|
n.stopInProc()
|
|
return err
|
|
}
|
|
// All API endpoints started successfully
|
|
n.rpcAPIs = apis
|
|
return nil
|
|
}
|
|
|
|
// startInProc initializes an in-process RPC endpoint.
|
|
func (n *Node) startInProc(apis []rpc.API) error {
|
|
// Register all the APIs exposed by the services
|
|
handler := rpc.NewServer()
|
|
for _, api := range apis {
|
|
if err := handler.RegisterName(api.Namespace, api.Service); err != nil {
|
|
return err
|
|
}
|
|
n.log.Debug("InProc registered", "namespace", api.Namespace)
|
|
}
|
|
n.inprocHandler = handler
|
|
return nil
|
|
}
|
|
|
|
// stopInProc terminates the in-process RPC endpoint.
|
|
func (n *Node) stopInProc() {
|
|
if n.inprocHandler != nil {
|
|
n.inprocHandler.Stop()
|
|
n.inprocHandler = nil
|
|
}
|
|
}
|
|
|
|
// startIPC initializes and starts the IPC RPC endpoint.
|
|
func (n *Node) startIPC(apis []rpc.API) error {
|
|
if n.ipcEndpoint == "" {
|
|
return nil // IPC disabled.
|
|
}
|
|
listener, handler, err := rpc.StartIPCEndpoint(n.ipcEndpoint, apis)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
n.ipcListener = listener
|
|
n.ipcHandler = handler
|
|
n.log.Info("IPC endpoint opened", "url", n.ipcEndpoint)
|
|
return nil
|
|
}
|
|
|
|
// stopIPC terminates the IPC RPC endpoint.
|
|
func (n *Node) stopIPC() {
|
|
if n.ipcListener != nil {
|
|
n.ipcListener.Close()
|
|
n.ipcListener = nil
|
|
|
|
n.log.Info("IPC endpoint closed", "url", n.ipcEndpoint)
|
|
}
|
|
if n.ipcHandler != nil {
|
|
n.ipcHandler.Stop()
|
|
n.ipcHandler = nil
|
|
}
|
|
}
|
|
|
|
// startHTTP initializes and starts the HTTP RPC endpoint.
|
|
func (n *Node) startHTTP(endpoint string, apis []rpc.API, modules []string, cors []string, vhosts []string, timeouts rpc.HTTPTimeouts) error {
|
|
// Short circuit if the HTTP endpoint isn't being exposed
|
|
if endpoint == "" {
|
|
return nil
|
|
}
|
|
listener, handler, err := rpc.StartHTTPEndpoint(endpoint, apis, modules, cors, vhosts, timeouts)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
n.log.Info("HTTP endpoint opened", "url", fmt.Sprintf("http://%s", endpoint), "cors", strings.Join(cors, ","), "vhosts", strings.Join(vhosts, ","))
|
|
// All listeners booted successfully
|
|
n.httpEndpoint = endpoint
|
|
n.httpListener = listener
|
|
n.httpHandler = handler
|
|
|
|
return nil
|
|
}
|
|
|
|
// stopHTTP terminates the HTTP RPC endpoint.
|
|
func (n *Node) stopHTTP() {
|
|
if n.httpListener != nil {
|
|
n.httpListener.Close()
|
|
n.httpListener = nil
|
|
|
|
n.log.Info("HTTP endpoint closed", "url", fmt.Sprintf("http://%s", n.httpEndpoint))
|
|
}
|
|
if n.httpHandler != nil {
|
|
n.httpHandler.Stop()
|
|
n.httpHandler = nil
|
|
}
|
|
}
|
|
|
|
// startWS initializes and starts the websocket RPC endpoint.
|
|
func (n *Node) startWS(endpoint string, apis []rpc.API, modules []string, wsOrigins []string, exposeAll bool) error {
|
|
// Short circuit if the WS endpoint isn't being exposed
|
|
if endpoint == "" {
|
|
return nil
|
|
}
|
|
listener, handler, err := rpc.StartWSEndpoint(endpoint, apis, modules, wsOrigins, exposeAll)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
n.log.Info("WebSocket endpoint opened", "url", fmt.Sprintf("ws://%s", listener.Addr()))
|
|
// All listeners booted successfully
|
|
n.wsEndpoint = endpoint
|
|
n.wsListener = listener
|
|
n.wsHandler = handler
|
|
|
|
return nil
|
|
}
|
|
|
|
// stopWS terminates the websocket RPC endpoint.
|
|
func (n *Node) stopWS() {
|
|
if n.wsListener != nil {
|
|
n.wsListener.Close()
|
|
n.wsListener = nil
|
|
|
|
n.log.Info("WebSocket endpoint closed", "url", fmt.Sprintf("ws://%s", n.wsEndpoint))
|
|
}
|
|
if n.wsHandler != nil {
|
|
n.wsHandler.Stop()
|
|
n.wsHandler = nil
|
|
}
|
|
}
|
|
|
|
// Stop terminates a running node along with all it's services. In the node was
|
|
// not started, an error is returned.
|
|
func (n *Node) Stop() error {
|
|
n.lock.Lock()
|
|
defer n.lock.Unlock()
|
|
|
|
// Short circuit if the node's not running
|
|
if n.server == nil {
|
|
return ErrNodeStopped
|
|
}
|
|
|
|
// Terminate the API, services and the p2p server.
|
|
n.stopWS()
|
|
n.stopHTTP()
|
|
n.stopIPC()
|
|
n.rpcAPIs = nil
|
|
failure := &StopError{
|
|
Services: make(map[reflect.Type]error),
|
|
}
|
|
for kind, service := range n.services {
|
|
if err := service.Stop(); err != nil {
|
|
failure.Services[kind] = err
|
|
}
|
|
}
|
|
n.server.Stop()
|
|
n.services = nil
|
|
n.server = nil
|
|
|
|
// Release instance directory lock.
|
|
if n.instanceDirLock != nil {
|
|
if err := n.instanceDirLock.Release(); err != nil {
|
|
n.log.Error("Can't release datadir lock", "err", err)
|
|
}
|
|
n.instanceDirLock = nil
|
|
}
|
|
|
|
// unblock n.Wait
|
|
close(n.stop)
|
|
|
|
// Remove the keystore if it was created ephemerally.
|
|
var keystoreErr error
|
|
if n.ephemeralKeystore != "" {
|
|
keystoreErr = os.RemoveAll(n.ephemeralKeystore)
|
|
}
|
|
|
|
if len(failure.Services) > 0 {
|
|
return failure
|
|
}
|
|
if keystoreErr != nil {
|
|
return keystoreErr
|
|
}
|
|
return nil
|
|
}
|
|
|
|
// Wait blocks the thread until the node is stopped. If the node is not running
|
|
// at the time of invocation, the method immediately returns.
|
|
func (n *Node) Wait() {
|
|
n.lock.RLock()
|
|
if n.server == nil {
|
|
n.lock.RUnlock()
|
|
return
|
|
}
|
|
stop := n.stop
|
|
n.lock.RUnlock()
|
|
|
|
<-stop
|
|
}
|
|
|
|
// Restart terminates a running node and boots up a new one in its place. If the
|
|
// node isn't running, an error is returned.
|
|
func (n *Node) Restart() error {
|
|
if err := n.Stop(); err != nil {
|
|
return err
|
|
}
|
|
if err := n.Start(); err != nil {
|
|
return err
|
|
}
|
|
return nil
|
|
}
|
|
|
|
// Attach creates an RPC client attached to an in-process API handler.
|
|
func (n *Node) Attach() (*rpc.Client, error) {
|
|
n.lock.RLock()
|
|
defer n.lock.RUnlock()
|
|
|
|
if n.server == nil {
|
|
return nil, ErrNodeStopped
|
|
}
|
|
return rpc.DialInProc(n.inprocHandler), nil
|
|
}
|
|
|
|
// RPCHandler returns the in-process RPC request handler.
|
|
func (n *Node) RPCHandler() (*rpc.Server, error) {
|
|
n.lock.RLock()
|
|
defer n.lock.RUnlock()
|
|
|
|
if n.inprocHandler == nil {
|
|
return nil, ErrNodeStopped
|
|
}
|
|
return n.inprocHandler, nil
|
|
}
|
|
|
|
// Server retrieves the currently running P2P network layer. This method is meant
|
|
// only to inspect fields of the currently running server, life cycle management
|
|
// should be left to this Node entity.
|
|
func (n *Node) Server() *p2p.Server {
|
|
n.lock.RLock()
|
|
defer n.lock.RUnlock()
|
|
|
|
return n.server
|
|
}
|
|
|
|
// Service retrieves a currently running service registered of a specific type.
|
|
func (n *Node) Service(service interface{}) error {
|
|
n.lock.RLock()
|
|
defer n.lock.RUnlock()
|
|
|
|
// Short circuit if the node's not running
|
|
if n.server == nil {
|
|
return ErrNodeStopped
|
|
}
|
|
// Otherwise try to find the service to return
|
|
element := reflect.ValueOf(service).Elem()
|
|
if running, ok := n.services[element.Type()]; ok {
|
|
element.Set(reflect.ValueOf(running))
|
|
return nil
|
|
}
|
|
return ErrServiceUnknown
|
|
}
|
|
|
|
// DataDir retrieves the current datadir used by the protocol stack.
|
|
// Deprecated: No files should be stored in this directory, use InstanceDir instead.
|
|
func (n *Node) DataDir() string {
|
|
return n.config.DataDir
|
|
}
|
|
|
|
// InstanceDir retrieves the instance directory used by the protocol stack.
|
|
func (n *Node) InstanceDir() string {
|
|
return n.config.instanceDir()
|
|
}
|
|
|
|
// AccountManager retrieves the account manager used by the protocol stack.
|
|
func (n *Node) AccountManager() *accounts.Manager {
|
|
return n.accman
|
|
}
|
|
|
|
// IPCEndpoint retrieves the current IPC endpoint used by the protocol stack.
|
|
func (n *Node) IPCEndpoint() string {
|
|
return n.ipcEndpoint
|
|
}
|
|
|
|
// HTTPEndpoint retrieves the current HTTP endpoint used by the protocol stack.
|
|
func (n *Node) HTTPEndpoint() string {
|
|
n.lock.Lock()
|
|
defer n.lock.Unlock()
|
|
|
|
if n.httpListener != nil {
|
|
return n.httpListener.Addr().String()
|
|
}
|
|
return n.httpEndpoint
|
|
}
|
|
|
|
// WSEndpoint retrieves the current WS endpoint used by the protocol stack.
|
|
func (n *Node) WSEndpoint() string {
|
|
n.lock.Lock()
|
|
defer n.lock.Unlock()
|
|
|
|
if n.wsListener != nil {
|
|
return n.wsListener.Addr().String()
|
|
}
|
|
return n.wsEndpoint
|
|
}
|
|
|
|
// EventMux retrieves the event multiplexer used by all the network services in
|
|
// the current protocol stack.
|
|
func (n *Node) EventMux() *event.TypeMux {
|
|
return n.eventmux
|
|
}
|
|
|
|
// OpenDatabase opens an existing database with the given name (or creates one if no
|
|
// previous can be found) from within the node's instance directory. If the node is
|
|
// ephemeral, a memory database is returned.
|
|
func (n *Node) OpenDatabase(name string, cache, handles int, namespace string) (ethdb.Database, error) {
|
|
if n.config.DataDir == "" {
|
|
return rawdb.NewMemoryDatabase(), nil
|
|
}
|
|
return rawdb.NewLevelDBDatabase(n.config.ResolvePath(name), cache, handles, namespace)
|
|
}
|
|
|
|
// OpenDatabaseWithFreezer opens an existing database with the given name (or
|
|
// creates one if no previous can be found) from within the node's data directory,
|
|
// also attaching a chain freezer to it that moves ancient chain data from the
|
|
// database to immutable append-only files. If the node is an ephemeral one, a
|
|
// memory database is returned.
|
|
func (n *Node) OpenDatabaseWithFreezer(name string, cache, handles int, freezer, namespace string) (ethdb.Database, error) {
|
|
if n.config.DataDir == "" {
|
|
return rawdb.NewMemoryDatabase(), nil
|
|
}
|
|
root := n.config.ResolvePath(name)
|
|
|
|
switch {
|
|
case freezer == "":
|
|
freezer = filepath.Join(root, "ancient")
|
|
case !filepath.IsAbs(freezer):
|
|
freezer = n.config.ResolvePath(freezer)
|
|
}
|
|
return rawdb.NewLevelDBDatabaseWithFreezer(root, cache, handles, freezer, namespace)
|
|
}
|
|
|
|
// ResolvePath returns the absolute path of a resource in the instance directory.
|
|
func (n *Node) ResolvePath(x string) string {
|
|
return n.config.ResolvePath(x)
|
|
}
|
|
|
|
// apis returns the collection of RPC descriptors this node offers.
|
|
func (n *Node) apis() []rpc.API {
|
|
return []rpc.API{
|
|
{
|
|
Namespace: "admin",
|
|
Version: "1.0",
|
|
Service: NewPrivateAdminAPI(n),
|
|
}, {
|
|
Namespace: "admin",
|
|
Version: "1.0",
|
|
Service: NewPublicAdminAPI(n),
|
|
Public: true,
|
|
}, {
|
|
Namespace: "debug",
|
|
Version: "1.0",
|
|
Service: debug.Handler,
|
|
}, {
|
|
Namespace: "web3",
|
|
Version: "1.0",
|
|
Service: NewPublicWeb3API(n),
|
|
Public: true,
|
|
},
|
|
}
|
|
}
|