libp2p-test-plans/multidim-interop
Marco Munizaga 3856de32fc
Revert "Add zig-libp2p to interop tester (#167)" (#170)
2023-04-20 08:33:21 +02:00
..
compose-spec Add fast multidimensional interop tests (#97) 2023-01-10 14:26:27 -08:00
helpers Cache images slightly differently (#165) 2023-04-11 11:29:32 -07:00
impl Revert "Add zig-libp2p to interop tester (#167)" (#170) 2023-04-20 08:33:21 +02:00
src fix(interop): rename `webrtc` to `webrtc-direct` (#166) 2023-04-19 13:48:07 +02:00
.gitignore Add fast multidimensional interop tests (#97) 2023-01-10 14:26:27 -08:00
Makefile Revert "Add zig-libp2p to interop tester (#167)" (#170) 2023-04-20 08:33:21 +02:00
README.md Cache images slightly differently (#165) 2023-04-11 11:29:32 -07:00
dockerBuildWrapper.sh Update Go & JS tests to conform to the multidim interop test spec. (#121) 2023-02-10 17:00:53 -08:00
package-lock.json Cache images slightly differently (#165) 2023-04-11 11:29:32 -07:00
package.json Cache images slightly differently (#165) 2023-04-11 11:29:32 -07:00
renderResults.ts Add fast multidimensional interop tests (#97) 2023-01-10 14:26:27 -08:00
testplans.ts Introduce composite action for running ping interop tests (#123) 2023-02-02 10:17:22 +11:00
tsconfig.json Introduce composite action for running ping interop tests (#123) 2023-02-02 10:17:22 +11:00
versions.ts Revert "Add zig-libp2p to interop tester (#167)" (#170) 2023-04-20 08:33:21 +02:00

README.md

Interoperability test

This tests that different libp2p implementations can communicate with each other on each of their supported capabilities.

Each version of libp2p is defined in versions.ts. There the version defines its capabilities along with the id of its container image.

This repo and tests adhere to these constraints:

  1. Be reproducible for a given commit.
  2. Caching is an optimization. Things should be fine without it.
  3. If we have a cache hit, be fast.

Test spec

The implementation is run in a container and is passed parameters via environment variables. The current parameters are:

Name Description Is Optional
transport The transport to use no
muxer The muxer to use no, except when transport is one of quic, quic-v1, webtransport
security The security channel to use no, except when transport is one of quic, quic-v1, webtransport
is_dialer Should you dial or listen no
ip IP address to bind the listener to yes, default to "0.0.0.0"
redis_addr A different address to connect to redis (default redis:6379) yes, default to the redis host on port 6379
test_timeout_seconds Control the timeout of test. yes, default to 180 seconds.

The test should do two different things depending on if it's the dialer or listener.

Dialer

The dialer should emit all diagnostic logs to stderr. Only the final JSON string result should be emitted to stdout.

  1. Connect to the Redis instance.
  2. Create a libp2p node as defined by the environment variables.
  3. Get the listener's address via Redis' BLPOP using the listenerAddr key.
  4. Record the current instant as handshakeStartInstant.
  5. Connect to the listener.
  6. Ping the listener, and record the round trip duration as pingRTT
  7. Record the duration since handshakeStartInstant. This is handshakePlusOneRTT.
  8. Print to stdout the JSON formatted string: {"handshakePlusOneRTTMillis": handshakePlusOneRTT, "pingRTTMilllis": pingRTT}. Durations should be printed in milliseconds as a float.
  9. Exit with a code zero.

On error, the dialer should return a non-zero exit code.

Listener

The listener should emit all diagnostic logs to stderr.

  1. Connect to the Redis instance.
  2. Create a libp2p node as defined by the environment variables.
  3. Publish the listener's address via Redis' RPUSH using the listenerAddr key.
  4. Sleep for the duration of test_timeout_seconds. The test runner will kill this process when the dialer finishes.
  5. If the timeout is hit, exit with a non-zero error code.

On error, the listener should return a non-zero exit code.

Caching

The caching strategy is opinionated in an attempt to make things simpler and faster. Here's how it works:

  1. We cache the result of image.json in each implementation folder.
  2. The cache key is derived from the hashes of the files in the implementation folder.
  3. When loading from cache, if we have a cache hit, we load the image into docker and create the image.json file. We then call make -o image.json to allow the implementation to build any extra things from cache (e.g. JS-libp2p builds browser images from the same base as node). If we have a cache miss, we simply call make and build from scratch.
  4. When we push the cache we use the cache-key along with the docker platform (arm64 vs x86_64).