2980863ebc
* feat: automate adding new go-libp2p versions to perf * fix: put add-new-impl-versions.yml on schedule |
||
---|---|---|
.. | ||
impl | ||
runner | ||
terraform | ||
Makefile | ||
README.md |
README.md
libp2p performance benchmarking
This project includes the following components:
terraform/
: a Terraform scripts to provision infrastructureimpl/
: implementations of the libp2p perf protocol running on top of e.g. go-libp2p, rust-libp2p or Go's std-library https stackrunner/
: a set of scripts building and running the above implementations on the above infrastructure, reporting the results inbenchmark-results.json
Benchmark results can be visualized with https://observablehq.com/@libp2p-workspace/performance-dashboard.
Running via GitHub Action
- Create a pull request with your changes on https://github.com/libp2p/test-plans/.
- Trigger GitHub Action for branch on https://github.com/libp2p/test-plans/actions/workflows/perf.yml (see Run workflow button).
- Wait for action run to finish and to push a commit to your branch.
- Visualize results on https://observablehq.com/@libp2p-workspace/performance-dashboard.
Running manually
Prerequisites
- Terraform 1.5.4 or later
- Node.js 18 or later
- an AWS IAM user
Provision infrastructure
- Save your public SSH key as the file
./terraform/modules/short_lived/files/perf.pub
; or generate a new key pair withmake ssh-keygen
and add it to your SSH agent withmake ssh-add
. cd terraform/configs/local
terraform init
terraform apply
CLIENT_IP=$(terraform output -raw client_ip)
SERVER_IP=$(terraform output -raw server_ip)
Notes
-
While running terraform you may encounter the following error:
Error: collecting instance settings: reading EC2 Launch Template versions: couldn't find resource │ │ with module.short_lived_server[0].aws_instance.perf, │ on ../../modules/short_lived/main.tf line 15, in resource "aws_instance" "perf": │ 15: resource "aws_instance" "perf" {
-
This implies that you haven't deployed the long-lived infrastructure on your AWS account. To do so along with each short-lived deployment, you can set TF_VAR
long_lived_enabled
env variable to default totrue
. Terraform should then spin up the long-lived resources that are required for the short-lived resources to be created. -
It's best to destroy the infrastructure after you're done with your testing, you can do that by running
terraform destroy
.
Build and run libp2p implementations
Given you have provisioned your infrastructure, you can now build and run the libp2p implementations on the AWS instances.
cd runner
npm ci
npm run start -- --client-public-ip $CLIENT_IP --server-public-ip $SERVER_IP
- Note: The default number of iterations that perf will run is 10; desired iterations can be set with the
--iterations <value>
option.
Deprovision infrastructure
cd terraform/configs/local
terraform destroy
Adding a new implementation or a new version
- Add the implementation to new subdirectory in
impl/*
.- For a new implementation, create a folder
impl/<your-implementation-name>/
e.g.go-libp2p
- For a new version of an existing implementation, create a folder
impl/<your-implementation-name>/<your-implementation-version>
. - In that folder include a
Makefile
that builds an executable and stores it next to theMakefile
under the nameperf
. - Requirements for the executable:
- Running as a libp2p-perf server:
- The perf server must not exit as it will be closed by the test runner.
- The executable must accept the command flag
--run-server
which indicates it's running as server.
- Running as a libp2p-perf client
- Given that perf is a client driven set of benchmarks, the performance will be measured by the client.
- Input via command line
--server-address
--transport
(seerunner/versions.ts
for possible variants)--upload-bytes
number of bytes to upload per stream.--download-bytes
number of bytes to download per stream.
- Output
- Logging MUST go to
stderr
. - Measurement output is printed to
stdout
as JSON in the form of:
Note that the measurement includes the time to (1) establish the connection, (2) upload the bytes and (3) download the bytes.{"latency": 0.246442851}
- Logging MUST go to
- Input via command line
- Given that perf is a client driven set of benchmarks, the performance will be measured by the client.
- Running as a libp2p-perf server:
- For a new implementation, create a folder
- For a new implementation, in
impl/Makefile
include your implementation in theall
target. - For a new version, reference version in
runner/src/versions.ts
.