|
@@ -6,7 +6,7 @@ use regex::Regex;
|
|
|
|
|
|
use crate::run::{BENCH_BASE_PATH, BENCH_DATA_PATH};
|
|
|
|
|
|
-fn json_from_file<P>(path: P) -> anyhow::Result<HashMap<String, f64>>
|
|
|
+fn json_from_file<P>(path: P) -> anyhow::Result<HashMap<String, String>>
|
|
|
where
|
|
|
P: Debug,
|
|
|
P: AsRef<Path>
|
|
@@ -26,16 +26,14 @@ where
|
|
|
};
|
|
|
let mut map = HashMap::new();
|
|
|
for (key, val) in json_obj.iter() {
|
|
|
- let val: f64 = match val {
|
|
|
+ let val: String = match val {
|
|
|
JsonValue::Number(n) => {
|
|
|
- f64::from(*n)
|
|
|
+ f64::from(*n).to_string()
|
|
|
},
|
|
|
JsonValue::String(_) | JsonValue::Short(_) => {
|
|
|
val.as_str()
|
|
|
.unwrap()
|
|
|
- .strip_suffix("%")
|
|
|
- .ok_or(anyhow!("Expected percentage String, got: {:?}", val))?
|
|
|
- .parse()?
|
|
|
+ .to_string()
|
|
|
},
|
|
|
_ => bail!("Expected json number or string, got: {:?}", val),
|
|
|
};
|
|
@@ -127,146 +125,208 @@ pub fn process(opts: Options) -> anyhow::Result<()> {
|
|
|
let mut path = PathBuf::new();
|
|
|
path.push(BENCH_BASE_PATH);
|
|
|
path.push(BENCH_DATA_PATH);
|
|
|
- path.push(opts.seed);
|
|
|
|
|
|
let zmap_stats_regex = Regex::new(r"^[^\(;]+(?:\(.+left\))?; send: [^\(;]+done \(([\d\.]+) ([KMG]?)p/s avg\); recv: [^\(;]+\(([\d\.]+) ([KMG]?)p/s avg\); drops: [^\(;]+\(([\d\.]+) ([KMG]?)p/s avg\); hitrate: [^;]+$")?;
|
|
|
|
|
|
- let header_row = [
|
|
|
- "type", "filter-type",
|
|
|
- "subnet_size", "hitrate", "bloom_filter_bits", "bloom_filter_hash_count", "zmap_scanrate",
|
|
|
+ let run_header_row = [
|
|
|
+ "run", "type", "filter-type",
|
|
|
+ "subnet_size", "dataset_index", "filter_bits", "bloom_filter_hash_count", "zmap_scanrate",
|
|
|
"bpf_run_time_total", "bpf_run_count", "bpf_memory_lock",
|
|
|
- "filter_intern_build_time", "filter_intern_write_time",
|
|
|
- "filter_extern_time_clock", "filter_extern_cpu_p", "filter_extern_kernel_secs", "filter_extern_user_secs",
|
|
|
"zmap_send_rate", "zmap_receive_rate", "zmap_drop_rate",
|
|
|
"false_positive_count", "false_negative_count"
|
|
|
];
|
|
|
- let mut data_rows = vec![];
|
|
|
- data_rows.push(header_row.map(str::to_string));
|
|
|
|
|
|
- for subnet_dir in path.read_dir().context(format!("Failed to read subnet dirs in path {:?}", &path))? {
|
|
|
- let subnet_dir = subnet_dir.context(format!("Failed to read file info on file in path {:?}", &path))?;
|
|
|
- if !subnet_dir
|
|
|
+ let build_header_row = [
|
|
|
+ "type", "filter-type",
|
|
|
+ "subnet_size", "dataset_index", "bloom_filter_bits", "bloom_filter_hash_count",
|
|
|
+ "ips_subnet", "ips_entries", "ips_hitrate",
|
|
|
+ "filter_intern_build_time", "filter_intern_write_time",
|
|
|
+ "filter_extern_time_clock", "filter_extern_cpu_p", "filter_extern_kernel_secs", "filter_extern_user_secs",
|
|
|
+ ];
|
|
|
+
|
|
|
+ let mut run_data_rows = vec![];
|
|
|
+ run_data_rows.push(run_header_row.map(str::to_string));
|
|
|
+
|
|
|
+ let mut build_data_rows = vec![];
|
|
|
+ build_data_rows.push(build_header_row.map(str::to_string));
|
|
|
+
|
|
|
+ let get_or_default = |map: &mut HashMap<String, String>, k: &str| map
|
|
|
+ .remove(k).unwrap_or(String::from("-1"));
|
|
|
+
|
|
|
+ for run_dir in path.read_dir().context(format!("Failed to read subnet dirs in path {:?}", &path))? {
|
|
|
+ let run_dir = run_dir.context(format!("Failed to read file info on file in path {:?}", &path))?;
|
|
|
+ if !run_dir
|
|
|
.file_type()
|
|
|
- .context(format!("Failed to read file info on file {:?}", subnet_dir.path()))?
|
|
|
+ .context(format!("Failed to read file info on file {:?}", run_dir.path()))?
|
|
|
.is_dir() {
|
|
|
- bail!("Expected dir at {:?}", subnet_dir.path())
|
|
|
+ bail!("Expected dir at {:?}", run_dir.path());
|
|
|
}
|
|
|
- let subnet = subnet_dir.file_name().into_string().map_err(|e| anyhow!(format!("{:?}", e)))?;
|
|
|
- for hitrate_dir in subnet_dir.path().read_dir().context(format!("Failed to read hitrate dirs in path {:?}", subnet_dir.path()))? {
|
|
|
- let hitrate_dir = hitrate_dir.context(format!("Failed to read file info on file in path {:?}",subnet_dir.path()))?;
|
|
|
- if !hitrate_dir
|
|
|
+ let run = run_dir.file_name().into_string().map_err(|e| anyhow!(format!("{:?}", e)))?;
|
|
|
+ let is_build = run == "build";
|
|
|
+
|
|
|
+ let seed_path = run_dir.path().join(opts.seed.to_owned());
|
|
|
+
|
|
|
+ for subnet_dir in seed_path.read_dir().context(format!("Failed to read subnet dirs in path {:?}", &path))? {
|
|
|
+ let subnet_dir = subnet_dir.context(format!("Failed to read file info on file in path {:?}", &path))?;
|
|
|
+ if !subnet_dir
|
|
|
.file_type()
|
|
|
- .context(format!("Failed to read file info on file {:?}", hitrate_dir.path()))?
|
|
|
+ .context(format!("Failed to read file info on file {:?}", subnet_dir.path()))?
|
|
|
.is_dir() {
|
|
|
- bail!("Expected dir at {:?}", hitrate_dir.path())
|
|
|
+ bail!("Expected dir at {:?}", subnet_dir.path())
|
|
|
}
|
|
|
- let hitrate = hitrate_dir.file_name().into_string().map_err(|e| anyhow!(format!("{:?}", e)))?;
|
|
|
- let in_ips = ips_from_file(hitrate_dir.path().join("ips.txt")).context(format!("Failed to read ips from {:?}/ips.txt", hitrate_dir.path()))?;
|
|
|
- for bloom_dir in hitrate_dir.path().read_dir().context(format!("Failed to read bloom dirs in path {:?}", hitrate_dir.path()))? {
|
|
|
- let bloom_dir = bloom_dir.context(format!("Failed to read file info on file in path {:?}", hitrate_dir.path()))?;
|
|
|
- if !bloom_dir
|
|
|
+ let subnet = subnet_dir.file_name().into_string().map_err(|e| anyhow!(format!("{:?}", e)))?;
|
|
|
+ for dataset_dir in subnet_dir.path().read_dir().context(format!("Failed to read hitrate dirs in path {:?}", subnet_dir.path()))? {
|
|
|
+ let dataset_dir = dataset_dir.context(format!("Failed to read file info on file in path {:?}",subnet_dir.path()))?;
|
|
|
+ if !dataset_dir
|
|
|
.file_type()
|
|
|
- .context(format!("Failed to read file info on file {:?}", bloom_dir.path()))?
|
|
|
+ .context(format!("Failed to read file info on file {:?}", dataset_dir.path()))?
|
|
|
.is_dir() {
|
|
|
- continue;
|
|
|
+ bail!("Expected dir at {:?}", dataset_dir.path())
|
|
|
}
|
|
|
- let bloom_folder_name = bloom_dir
|
|
|
- .file_name()
|
|
|
- .into_string()
|
|
|
- .map_err(|e| anyhow!(format!("{:?}", e)))?;
|
|
|
- let (test_type, filter_type, bloom_bits, bloom_hashes) = if bloom_folder_name.contains('-') {
|
|
|
- let (bloom_bits, rem) = bloom_folder_name.split_once("-")
|
|
|
- .ok_or(anyhow!("Expected filename with -, got {:?}", bloom_dir.file_name()))?;
|
|
|
-
|
|
|
- let (filter_type, rem) = rem.split_once("-").unwrap_or((rem, ""));
|
|
|
-
|
|
|
- let (bloom_hashes, bpf_enabled) = if filter_type == "bloom" {
|
|
|
- rem.split_once("-").map(|(c, rem)| (c,rem=="bpf")).unwrap_or((rem, false))
|
|
|
- } else {
|
|
|
- ("0", rem == "bpf")
|
|
|
- };
|
|
|
-
|
|
|
- let bloom_bits = bloom_bits.to_string();
|
|
|
- let bloom_hashes = bloom_hashes.to_string();
|
|
|
- let test_type = if bpf_enabled {
|
|
|
- "bpf-stats"
|
|
|
- } else {
|
|
|
- "normal"
|
|
|
- };
|
|
|
- (test_type, filter_type, bloom_bits, bloom_hashes)
|
|
|
+ let dataset_index = dataset_dir.file_name().into_string().map_err(|e| anyhow!(format!("{:?}", e)))?;
|
|
|
+ let mut in_ips_path = path.clone();
|
|
|
+ in_ips_path.push("build");
|
|
|
+ in_ips_path.push(&opts.seed);
|
|
|
+ in_ips_path.push(&subnet);
|
|
|
+ in_ips_path.push(&dataset_index);
|
|
|
+ in_ips_path.push("ips.txt");
|
|
|
+
|
|
|
+ let in_ips = ips_from_file(&in_ips_path).context(format!("Failed to read ips from {:?}", &in_ips_path))?;
|
|
|
+ let ips_info = if is_build {
|
|
|
+ Some(json_from_file(dataset_dir.path().join("ips-info.json")).context(format!("Failed to read ip info from {:?}/ips-info.json", dataset_dir.path()))?)
|
|
|
} else {
|
|
|
- ("baseline","none", String::from("-1"), String::from("-1"))
|
|
|
+ None
|
|
|
};
|
|
|
-
|
|
|
- let bloom_path = bloom_dir.path();
|
|
|
-
|
|
|
- let mut filter_intern_time = json_from_file(bloom_path.join("filter_intern_time.json"))
|
|
|
- .context(format!("Failed to parse filter_intern_time.json for {:?}", bloom_path))?;
|
|
|
- let mut filter_extern_time = json_from_file(bloom_path.join("filter_extern_time.json"))
|
|
|
- .context(format!("Failed to parse filter_extern_time.json for {:?}", bloom_path))?;
|
|
|
-
|
|
|
- for scan_rate_dir in bloom_dir.path().read_dir().context(format!("Failed to read scan rate dirs in path {:?}", bloom_dir.path()))? {
|
|
|
- let scan_rate_dir = scan_rate_dir.context(format!("Failed to read file info on file in path {:?}", bloom_dir.path()))?;
|
|
|
- if !scan_rate_dir
|
|
|
+ for filter_dir in dataset_dir.path().read_dir().context(format!("Failed to read bloom dirs in path {:?}", dataset_dir.path()))? {
|
|
|
+ let filter_dir = filter_dir.context(format!("Failed to read file info on file in path {:?}", dataset_dir.path()))?;
|
|
|
+ if !filter_dir
|
|
|
.file_type()
|
|
|
- .context(format!("Failed to read file info on file {:?}", scan_rate_dir.path()))?
|
|
|
+ .context(format!("Failed to read file info on file {:?}", filter_dir.path()))?
|
|
|
.is_dir() {
|
|
|
continue;
|
|
|
}
|
|
|
- let scan_rate = scan_rate_dir.file_name().to_str().unwrap().to_string();
|
|
|
-
|
|
|
- let wd_path = scan_rate_dir.path();
|
|
|
- let mut bpf_stats = json_from_file(wd_path.join("bpf_stats.json"))
|
|
|
- .context(format!("Failed to parse bpf_stats.json for {:?}", wd_path))?;
|
|
|
- let out_ips = ips_from_file(wd_path.join("zmap_out_ips.txt"))
|
|
|
- .context(format!("Failed to parse zmap_out_ips.txt from {:?}", wd_path))?;
|
|
|
- let zmap_stats = zmap_stats(wd_path.join("zmap_stats.txt"), &zmap_stats_regex)
|
|
|
- .context(format!("Failed to parse zmap_stats.txt from {:?}", wd_path))?;
|
|
|
-
|
|
|
- let get_or_default = |map: &mut HashMap<String, f64>, k: &str| map
|
|
|
- .get(k).unwrap_or(&-1f64).to_string();
|
|
|
-
|
|
|
- let data_row = (|| {
|
|
|
- Ok([
|
|
|
- test_type.to_owned(),
|
|
|
- filter_type.to_owned(),
|
|
|
- subnet.clone(), hitrate.clone(), bloom_bits.clone(), bloom_hashes.clone(), scan_rate.clone(),
|
|
|
- get_or_default(&mut bpf_stats, "run_time"),
|
|
|
- get_or_default(&mut bpf_stats, "run_count"),
|
|
|
- get_or_default(&mut bpf_stats, "mem_lock"),
|
|
|
- get_or_default(&mut filter_intern_time, "build"),
|
|
|
- get_or_default(&mut filter_intern_time, "write"),
|
|
|
- get_or_default(&mut filter_extern_time, "clock"),
|
|
|
- get_or_default(&mut filter_extern_time, "cpu_p"),
|
|
|
- get_or_default(&mut filter_extern_time, "kernel_s"),
|
|
|
- get_or_default(&mut filter_extern_time, "user_s"),
|
|
|
- zmap_stats.0.to_string(),
|
|
|
- zmap_stats.1.to_string(),
|
|
|
- zmap_stats.2.to_string(),
|
|
|
- out_ips.difference(&in_ips).count().to_string(),
|
|
|
- in_ips.difference(&out_ips).count().to_string(),
|
|
|
- ])
|
|
|
- })().map_err(|key: String| {
|
|
|
- anyhow!("Failed to read data point {} for {:?}", key, wd_path)
|
|
|
- })?;
|
|
|
-
|
|
|
- let mut f = File::create(wd_path.join("data_row.csv"))?;
|
|
|
- f.write_all(header_row.join(",").as_bytes())?;
|
|
|
- f.write(&[b'\n'])?;
|
|
|
- f.write_all(data_row.join(",").as_bytes())?;
|
|
|
- data_rows.push(data_row);
|
|
|
+ let filter_folder_name = filter_dir
|
|
|
+ .file_name()
|
|
|
+ .into_string()
|
|
|
+ .map_err(|e| anyhow!(format!("{:?}", e)))?;
|
|
|
+ let (test_type, filter_type, bits, bloom_hashes) = if filter_folder_name.contains('-') {
|
|
|
+ let (bits, rem) = filter_folder_name.split_once("-")
|
|
|
+ .ok_or(anyhow!("Expected filename with -, got {:?}", filter_dir.file_name()))?;
|
|
|
+
|
|
|
+ let (filter_type, rem) = rem.split_once("-").unwrap_or((rem, ""));
|
|
|
+
|
|
|
+ let (bloom_hashes, bpf_enabled) = if filter_type == "bloom" {
|
|
|
+ rem.split_once("-").map(|(c, rem)| (c,rem=="bpf")).unwrap_or((rem, false))
|
|
|
+ } else {
|
|
|
+ ("0", rem == "bpf")
|
|
|
+ };
|
|
|
+
|
|
|
+ let bits = bits.to_string();
|
|
|
+ let bloom_hashes = bloom_hashes.to_string();
|
|
|
+ let test_type = if bpf_enabled {
|
|
|
+ "bpf-stats"
|
|
|
+ } else {
|
|
|
+ "normal"
|
|
|
+ };
|
|
|
+ (test_type, filter_type, bits, bloom_hashes)
|
|
|
+ } else {
|
|
|
+ ("baseline","none", String::from("-1"), String::from("-1"))
|
|
|
+ };
|
|
|
+
|
|
|
+ if is_build {
|
|
|
+ let bloom_path = filter_dir.path();
|
|
|
+ let mut ips_info = ips_info.clone().unwrap();
|
|
|
+
|
|
|
+ let mut filter_intern_time = json_from_file(bloom_path.join("filter_intern_time.json"))
|
|
|
+ .context(format!("Failed to parse filter_intern_time.json for {:?}", bloom_path))?;
|
|
|
+ let mut filter_extern_time = json_from_file(bloom_path.join("filter_extern_time.json"))
|
|
|
+ .context(format!("Failed to parse filter_extern_time.json for {:?}", bloom_path))?;
|
|
|
+
|
|
|
+ let build_data_row = (|| {
|
|
|
+ Ok([
|
|
|
+ test_type.to_owned(),
|
|
|
+ filter_type.to_owned(),
|
|
|
+ subnet.clone(), dataset_index.clone(), bits.clone(), bloom_hashes.clone(),
|
|
|
+ get_or_default(&mut ips_info, "subnet"),
|
|
|
+ get_or_default(&mut ips_info, "entries"),
|
|
|
+ get_or_default(&mut ips_info, "hitrate"),
|
|
|
+ get_or_default(&mut filter_intern_time, "build"),
|
|
|
+ get_or_default(&mut filter_intern_time, "write"),
|
|
|
+ get_or_default(&mut filter_extern_time, "clock"),
|
|
|
+ get_or_default(&mut filter_extern_time, "cpu_p"),
|
|
|
+ get_or_default(&mut filter_extern_time, "kernel_s"),
|
|
|
+ get_or_default(&mut filter_extern_time, "user_s"),
|
|
|
+ ])
|
|
|
+ })().map_err(|key: String| {
|
|
|
+ anyhow!("Failed to read build data point {} for {:?}", key, bloom_path)
|
|
|
+ })?;
|
|
|
+
|
|
|
+ let mut f = File::create(bloom_path.join("data_row.csv"))?;
|
|
|
+ f.write_all(build_header_row.join(",").as_bytes())?;
|
|
|
+ f.write(&[b'\n'])?;
|
|
|
+ f.write_all(build_data_row.join(",").as_bytes())?;
|
|
|
+ build_data_rows.push(build_data_row);
|
|
|
+ } else {
|
|
|
+ for scan_rate_dir in filter_dir.path().read_dir().context(format!("Failed to read scan rate dirs in path {:?}", filter_dir.path()))? {
|
|
|
+ let scan_rate_dir = scan_rate_dir.context(format!("Failed to read file info on file in path {:?}", filter_dir.path()))?;
|
|
|
+ if !scan_rate_dir
|
|
|
+ .file_type()
|
|
|
+ .context(format!("Failed to read file info on file {:?}", scan_rate_dir.path()))?
|
|
|
+ .is_dir() {
|
|
|
+ continue;
|
|
|
+ }
|
|
|
+ let scan_rate = scan_rate_dir.file_name().to_str().unwrap().to_string();
|
|
|
+
|
|
|
+ let wd_path = scan_rate_dir.path();
|
|
|
+ let mut bpf_stats = json_from_file(wd_path.join("bpf_stats.json"))
|
|
|
+ .context(format!("Failed to parse bpf_stats.json for {:?}", wd_path))?;
|
|
|
+ let out_ips = ips_from_file(wd_path.join("zmap_out_ips.txt"))
|
|
|
+ .context(format!("Failed to parse zmap_out_ips.txt from {:?}", wd_path))?;
|
|
|
+ let zmap_stats = zmap_stats(wd_path.join("zmap_stats.txt"), &zmap_stats_regex)
|
|
|
+ .context(format!("Failed to parse zmap_stats.txt from {:?}", wd_path))?;
|
|
|
+
|
|
|
+ let run_data_row = (|| {
|
|
|
+ Ok([
|
|
|
+ run.to_owned(),
|
|
|
+ test_type.to_owned(),
|
|
|
+ filter_type.to_owned(),
|
|
|
+ subnet.clone(), dataset_index.clone(), bits.clone(), bloom_hashes.clone(), scan_rate.clone(),
|
|
|
+ get_or_default(&mut bpf_stats, "run_time"),
|
|
|
+ get_or_default(&mut bpf_stats, "run_count"),
|
|
|
+ get_or_default(&mut bpf_stats, "mem_lock"),
|
|
|
+ zmap_stats.0.to_string(),
|
|
|
+ zmap_stats.1.to_string(),
|
|
|
+ zmap_stats.2.to_string(),
|
|
|
+ out_ips.difference(&in_ips).count().to_string(),
|
|
|
+ in_ips.difference(&out_ips).count().to_string(),
|
|
|
+ ])
|
|
|
+ })().map_err(|key: String| {
|
|
|
+ anyhow!("Failed to read data point {} for {:?}", key, wd_path)
|
|
|
+ })?;
|
|
|
+
|
|
|
+ let mut f = File::create(wd_path.join("data_row.csv"))?;
|
|
|
+ f.write_all(run_header_row.join(",").as_bytes())?;
|
|
|
+ f.write(&[b'\n'])?;
|
|
|
+ f.write_all(run_data_row.join(",").as_bytes())?;
|
|
|
+ run_data_rows.push(run_data_row);
|
|
|
|
|
|
+ }
|
|
|
+ }
|
|
|
}
|
|
|
}
|
|
|
}
|
|
|
}
|
|
|
|
|
|
- let data = data_rows.into_iter()
|
|
|
- .map(|row| row.join(","))
|
|
|
- .fold(String::new(), |a, b| a + &b + "\n");
|
|
|
+ let run_data = run_data_rows.into_iter()
|
|
|
+ .map(|row| row.join(","))
|
|
|
+ .fold(String::new(), |a, b| a + &b + "\n");
|
|
|
|
|
|
+ let build_data = build_data_rows.into_iter()
|
|
|
+ .map(|row| row.join(","))
|
|
|
+ .fold(String::new(), |a, b| a + &b + "\n");
|
|
|
|
|
|
- File::create(path.join("data.csv"))?.write_all(data.as_bytes())?;
|
|
|
+ File::create(path.join("run_data.csv"))?.write_all(run_data.as_bytes())?;
|
|
|
+ File::create(path.join("build_data.csv"))?.write_all(build_data.as_bytes())?;
|
|
|
|
|
|
|
|
|
Ok(())
|