mirror of
https://github.com/logos-storage/bittorrent-benchmarks.git
synced 2026-01-02 21:13:11 +00:00
180 lines
5.1 KiB
R
180 lines
5.1 KiB
R
PIECE_SIZE <- 262144
|
|
|
|
piece_count <- function(experiment) {
|
|
experiment$meta$file_size / PIECE_SIZE
|
|
}
|
|
|
|
extract_repetitions <- function(deluge_torrent_download) {
|
|
deluge_torrent_download |>
|
|
mutate(
|
|
temp = str_remove(torrent_name, '^dataset-'),
|
|
seed_set = as.numeric(str_extract(temp, '^\\d+')),
|
|
run = as.numeric(str_extract(temp, '\\d+$'))
|
|
) |>
|
|
rename(piece = value) |>
|
|
select(-temp, -name)
|
|
}
|
|
|
|
compute_pieces <- function(deluge_torrent_download, n_pieces) {
|
|
deluge_torrent_download |>
|
|
group_by(node, seed_set, run) |>
|
|
arrange(timestamp) |>
|
|
mutate(
|
|
piece_count = seq_along(timestamp)
|
|
) |>
|
|
ungroup() |>
|
|
mutate(completed = piece_count / n_pieces)
|
|
}
|
|
|
|
check_incomplete_downloads <- function(deluge_torrent_download, n_pieces) {
|
|
incomplete_downloads <- deluge_torrent_download |>
|
|
group_by(node, seed_set, run) |>
|
|
count() |>
|
|
ungroup() |>
|
|
filter(n != n_pieces)
|
|
|
|
nrow(incomplete_downloads) == 0
|
|
}
|
|
|
|
check_mismatching_repetitions <- function(deluge_torrent_download, repetitions) {
|
|
mismatching_repetitions <- deluge_torrent_download |>
|
|
select(seed_set, node, run) |>
|
|
distinct() |>
|
|
group_by(seed_set, node) |>
|
|
count() |>
|
|
filter(n != repetitions)
|
|
|
|
nrow(mismatching_repetitions) == 0
|
|
}
|
|
|
|
compute_download_times <- function(meta, request_event, deluge_torrent_download, group_id) {
|
|
n_leechers <- meta$nodes$network_size - meta$seeders
|
|
|
|
download_start <- request_event |>
|
|
select(-request_id) |>
|
|
filter(name == 'leech', type == 'RequestEventType.start') |>
|
|
mutate(
|
|
# We didn't log those on the runner side so I have to reconstruct them.
|
|
run = rep(rep(
|
|
1:meta$repetitions - 1,
|
|
each = n_leechers), times=meta$seeder_sets),
|
|
seed_set = rep(
|
|
1:meta$seeder_sets - 1,
|
|
each = n_leechers * meta$repetitions),
|
|
) |>
|
|
transmute(node = destination, run, seed_set, seed_request_time = timestamp)
|
|
|
|
download_times <- deluge_torrent_download |>
|
|
left_join(download_start, by = c('node', 'run', 'seed_set')) |>
|
|
mutate(
|
|
elapsed_download_time = as.numeric(timestamp - seed_request_time)
|
|
) |>
|
|
group_by(node, run, seed_set) |>
|
|
mutate(
|
|
time_to_first_byte = min(timestamp),
|
|
lookup_time = as.numeric(time_to_first_byte - seed_request_time)
|
|
) |>
|
|
ungroup()
|
|
|
|
if (nrow(download_times |>
|
|
filter(elapsed_download_time < 0 | lookup_time < 0)) > 0) {
|
|
stop('Calculation for download times contains negative numbers')
|
|
}
|
|
|
|
download_times
|
|
}
|
|
|
|
check_seeder_count <- function(download_times, seeders) {
|
|
mismatching_seeders <- download_times |>
|
|
filter(is.na(seed_request_time)) |>
|
|
select(node, seed_set, run) |>
|
|
distinct() |>
|
|
group_by(seed_set, run) |>
|
|
count() |>
|
|
filter(n != seeders)
|
|
|
|
nrow(mismatching_seeders) == 0
|
|
}
|
|
|
|
download_stats <- function(download_times) {
|
|
download_times |>
|
|
filter(!is.na(elapsed_download_time)) |>
|
|
group_by(piece_count, completed) |>
|
|
summarise(
|
|
mean = mean(elapsed_download_time),
|
|
median = median(elapsed_download_time),
|
|
max = max(elapsed_download_time),
|
|
min = min(elapsed_download_time),
|
|
p90 = quantile(elapsed_download_time, p = 0.95),
|
|
p10 = quantile(elapsed_download_time, p = 0.05),
|
|
.groups = 'drop'
|
|
)
|
|
}
|
|
|
|
completion_time_stats <- function(download_times) {
|
|
completion_times <- download_times |>
|
|
filter(!is.na(elapsed_download_time)) |>
|
|
pull(elapsed_download_time)
|
|
|
|
|
|
tibble(
|
|
min = min(completion_times),
|
|
p05 = quantile(completion_times, p = 0.05),
|
|
p10 = quantile(completion_times, p = 0.10),
|
|
p20 = quantile(completion_times, p = 0.20),
|
|
median = median(completion_times),
|
|
p80 = quantile(completion_times, p = 0.80),
|
|
p90 = quantile(completion_times, p = 0.90),
|
|
p95 = quantile(completion_times, p = 0.95),
|
|
max = max(completion_times)
|
|
)
|
|
}
|
|
|
|
download_times <- function(experiment) {
|
|
meta <- experiment$meta
|
|
pieces <- experiment |> piece_count()
|
|
downloads <- experiment$deluge_torrent_download |>
|
|
extract_repetitions() |>
|
|
compute_pieces(pieces)
|
|
|
|
if (!check_incomplete_downloads(downloads, pieces)) {
|
|
warning(glue::glue('Discard experiment {experiment$experiment_id} ',
|
|
'due to incomplete downloads'))
|
|
return(NULL)
|
|
}
|
|
|
|
if (!check_mismatching_repetitions(downloads, meta$repetitions)) {
|
|
warning(glue::glue('Discard experiment {experiment$experiment_id} ',
|
|
'due to mismatching repetitions'))
|
|
return(NULL)
|
|
}
|
|
|
|
download_times <- compute_download_times(
|
|
meta,
|
|
experiment$request_event,
|
|
downloads,
|
|
group_id
|
|
)
|
|
|
|
if (!check_seeder_count(download_times, meta$seeders)) {
|
|
warning(glue::glue('Undefined download times do not match seeder count'))
|
|
return(NULL)
|
|
}
|
|
|
|
download_times
|
|
}
|
|
|
|
|
|
compute_compact_summary <- function(download_ecdf) {
|
|
lapply(c(0.05, 0.5, 0.95), function(p)
|
|
download_ecdf |>
|
|
filter(completed >= p) |>
|
|
slice_min(completed)
|
|
) |>
|
|
bind_rows() |>
|
|
select(completed, network_size, file_size, seeders, leechers, median) |>
|
|
pivot_wider(id_cols = c('file_size', 'network_size', 'seeders', 'leechers'),
|
|
names_from = completed, values_from = median)
|
|
}
|
|
|