⚠️ As of September 27, 2023 the Dreamboat code base is no longer being actively maintained by the Blocknative team. Please see this post for more context.
An Ethereum 2.0 Relay for proposer-builder separation (PBS) with MEV-boost. Commissioned, built, and put to sea by Blocknative.
Install with go get
, or download the latest release here.
$ go install github.com/blocknative/dreamboat
$ dreamboat --help
Documentation can be found here.
Dreamboat is a relay for Ethereum 2.0, it ships blocks from builders to validators. Dreamboat is for anyone who wants to strengthen the Eth network by running their own relay. Dreamboat's all-in-one design, modular code base and stunning performance will dazzle beginners and seasoned hackers alike. All aboard!
More precisely, Dreamboat is a spec-compliant builder and relay that allows block-builders and validators to interface with each other through MEV-boost.
Eventually Dreamboat will vanish, as if a dream, once in-protocol PBS is live. Until then, it sails the ephemeral seas of block space, delivering its trusted cargo of blocks to validators, guided simply by a lighthouse.
Why should you opt for Dreamboat instead of other relays?
Unlike centralized relays that heavily rely on a single centralized source of truth, such as Redis, for bid propagation, the Relay project takes a distributed approach. We deliver bids to distributed instances, allowing for improved scalability, fault tolerance, and reduced reliance on a single point of failure. This decentralized bid delivery mechanism ensures higher availability and robustness of the relay network.
One of the key differentiators of the Relay project is its heavy utilization of memory for serving various operations. Instead of relying on frequent database access, we leverage in-memory storage to serve most of the required data. This approach results in significantly faster response times and reduced latency for bid retrieval, payload retrieval, and other essential operations.
In contrast to relying solely on a costly PostgreSQL database for storing large volumes of data, the Relay project adopts a more cost-effective approach. We store increasing amounts of data in files, utilizing file-based storage systems. This strategy not only reduces infrastructure costs but also enhances performance and scalability, as file operations can be optimized and scaled more efficiently.
Relay diversity plays a crucial role in ensuring the resilience, security, and efficiency of the overall network. By running relays with different configurations and settings, we can achieve a more diverse and robust network architecture.
In addition to configuring the relay using the config.ini
file (explained below), you can also use command-line flags to customize its behavior. The following flags are available:
-loglvl
: Sets the logging level for the relay. The default value isinfo
. Possible options includetrace
,debug
,info
,warn
,error
, orfatal
.-logfmt
: Specifies the log format for the relay. The default value istext
. Possible options includetext
,json
, ornone
.-config
: Specifies the path to the configuration file needed for the relay to run.-datadir
: Sets the data directory where blocks and validators are stored in the default datastore implementation. The default value is/tmp/relay
.
The following flags control specific relay functionalities:
-relay-distribution
: Enables running the relay as a distributed system with multiple replicas. By default, this flag is set tofalse
.-warehouse
: Enables warehouse storage of data. By default, this flag is set totrue
.
The relay system provides various configuration options that can be customized through the config.ini
file, specified through the -config
flag. Each section and its respective items are explained below, along with their default values:
-
[external_http]
:address
(default: "0.0.0.0:18550"): The IP address and port on which the relay serves external connections.read_timeout
(default: 5s): The maximum duration for reading the entire request, including the body.idle_timeout
(default: 5s): The maximum amount of time to wait for the next request when keep-alives are enabled.write_timeout
(default: 5s): The maximum duration before timing out writes of the response.
-
[internal_http]
:address
(default: "0.0.0.0:19550"): The IP address and port on which the internal metrics profiling and management server operates.read_timeout
(default: 5s): The maximum duration for reading the entire request, including the body.idle_timeout
(default: 5s): The maximum amount of time to wait for the next request when keep-alives are enabled.write_timeout
(default: 5s): The maximum duration before timing out writes of the response.
-
[api]
:allowed_builders
(default: []): A comma-separated list of allowed builder public keys.submission_limit_rate
(default: 2): The rate limit for submission requests (per second).submission_limit_burst
(default: 2): The burst value for the submission request rate limit.limitter_cache_size
(default: 1000): The size of the rate limiter cache entries.data_limit
(default: 450): The limit of data (in bytes) returned in one response.errors_on_disable
(default: false): A flag indicating whether to return errors when the endpoint is disabled.
-
[relay]
:network
(default: "mainnet"): The name of the network in which the relay operates.secret_key
(default: ""): The secret key used to sign messages.publish_block
(default: true): A flag indicating whether to publish payloads to beacon nodes after delivery.get_payload_response_delay
(default: 800ms): The delay for responding to the GetPayload request.get_payload_request_time_limit
(default: 4s): The deadline for calling GetPayload.allowed_builders
(default: []): A comma-separated list of allowed builder public keys.
-
[beacon]
:addresses
(default: []): A comma-separated list of URLs to beacon endpoints.publish_addresses
(default: []): A comma-separated list of URLs to beacon endpoints for publishing.payload_attributes_subscription
(default: true): A flag indicating whether payload attributes should be enabled.event_timeout
(default: 16s): The timeout for beacon events.event_restart
(default: 5): The number of times to restart beacon events.query_timeout
(default: 20s): The timeout for beacon queries.
-
[verify]
:workers
(default: 2000): The number of workers running verify in parallel.queue_size
(default: 100000): The size of the verify queue.
-
[validators]
:db.url
(default: ""): The database connection query string.db.max_open_conns
(default: 10): The maximum number of open connections to the database.db.max_idle_conns
(default: 10): The maximum number of idle connections in the connection pool.db.conn_max_idle_time
(default: 15s): The maximum amount of time a connection can remain idle.badger.ttl
(default: 24h): The time-to-live duration for BadgerDB.queue_size
(default: 100000): The size of the response queue.store_workers
(default: 400): The number of workers storing validators in parallel.registrations_cache_size
(default: 600000): The size of the registrations cache.registrations_cache_ttl
(default: 1h): The time-to-live duration for the registrations read cache.registrations_write_cache_ttl
(default: 12h): The time-to-live duration for the registrations write cache.
-
[block_simulation]
:rpc.address
(default: ""): The block validation raw URL (e.g., ipc path).ws.address
(default: []): A comma-separated list of block validation endpoint addresses.ws.retry
(default: true): Retry to other WebSocket connections on failure.http.address
(default: ""): The block validation address.
-
[payload]
:badger.ttl
(default: 24h): The time-to-live duration for BadgerDB.cache_size
(default: 1000): The number of payloads to cache for fast in-memory reads.redis.read.address
(default: ""): The read Redis address.redis.write.address
(default: ""): The write Redis address.
-
[dataapi]
:db.url
(default: ""): The database connection query string.badger.ttl
(default: 24h): The time-to-live duration for BadgerDB.
-
[warehouse]
:directory
(default: "/data/relay/warehouse"): The data directory where the data is stored in the warehouse.workers
(default: 32): The number of workers for storing data in the warehouse. If set to 0, data is not exported.buffer
(default: 1000): The size of the buffer for processing requests.
-
[distributed]
:redis.topic
(default: "relay"): The Redis stream topic.redis.address
(default: ""): The Redis stream address.id
(default: ""): The distributed instance ID.workers
(default: 100): The number of workers for storing data in the warehouse. If set to 0, data is not exported.stream_queue_size
(default: 100): The stream internal channel size.stream_served_bids
(default: true): Stream the entire block for every bid served in GetHeader requests.
Please note that these values can be modified in the config.ini
file according to your specific requirements.
Do you need to make your relay work with a different blockchain that is not:
- Mainnet
- Kiln
- Ropsten
- Sepolia
- Goerli
Then you need to extend the relay by creating a json file called networks.json
, and put it inside datadir
(datadir
is a parameter you set when running the relay).
Inside the network.json file, you need to specify the network name and three other information:
- GenesisForkVersion
- GenesisValidatorsRoot
- BellatrixForkVersion
Here is an example of how you should format the networks:
{
"blocknative":{
"GenesisForkVersion": "",
"GenesisValidatorsRoot":"",
"BellatrixForkVersion": ""
},
"myNetwork":{
"GenesisForkVersion": "",
"GenesisValidatorsRoot":"",
"BellatrixForkVersion": ""
},
}
This is a from-scratch implementation of a PBS relay–aimed at strengthening #RelayDiversity–with the following design goals:
- Extensibility. Dreamboat provides well-factored interfaces that make it easy to change its behavior, including the database, beacon connection and even the relay core logic.
- Performance. We have put significant engineering effort into tuning Dreamboat's performance, and have much, much more in store. Concurrency is carefully tuned, we cache as much as we can, and we try to serve requests from memory. The result is a ship that doesn't just sail... it glides.
- Transparency. Dreamboat implements the Relay Data API, so you can audit past proposals. But we take a much broader view of transparency, which includes code transparency. A major cleanup effort is underway to make–and keep–Dreamboat's code legible. We believe this is an essential part of ensuring transparency and reliability.
We are continuously improving Dreamoat's runtime performance, standards compliance, reliability and transparency. We would also like to thank the Flashbots team for their open-source tooling, which helped us get Dreamboat up and running in short order, and for their thoughtful comments on implementation.
Stuck? Don't despair! Drop by our discord to ask for help, or just to say hello. We're friendly! đź‘‹
$ go run cmd/dreamboat
Proposer Builder Separation is a solution to shield validators from the centralizing force the black hole of MEV has released. The next chapter in this saga is to protect builders from too extreme of centralizing forces.
Dreamboat follows semantic versioning. Changes to the REST API are governed by the Flashbots specifications.
The main
branch is NOT guaranteed to be stable. Use the latest git tag in production!