-
Notifications
You must be signed in to change notification settings - Fork 0
Performance Benchmarks Journal
Unless otherwise specified, all runs were relative to a Mac M1 platform, which is a 4+4 core platform
Danilo Lessa Bernardineli (BlockScience)
For varying N_sweep_samples (N_t=2, N_s=2):
- n=1: 0.09s (212= 4 state measurements, 44P/s)
- n=10: 0.52s (2102= 40 state measurements, 77P/s)
- n=100: 5.85s (21002= 400 state measurements, 68P/s)
- n=200: 12.03s (22002= 800 state measurements, 60P/s)
- n=400: 32.55s (24002= 1600 state measurements, 50P/s)
- n=600: error
- n=1000: error
For varying N_t with N_sweep_samples=10, N_s=2:
- N_t=2: 0.51s (2102= 40 state measurements, 78P/s)
- N_t=20: 2.47s (2102= 400 state measurements, 162P/s)
- N_t=200: 42.72s (2102= 4000 state measurements, 93P/s)
For varying N_t with N_sw=1, N_s=10
- N_t=10: 0.7s (10110= 100 state measurements, 142P/s)
- N_t=100: 6.75s (101100 = 1,000 state measurements, 148P/s)
- N_t=1000: 283.0s (1011000 = 10,000 state measurements, 35P/s)
Last run: N_sw=20, N_s=2, N_t=100: 20.3s (220100= 100 state measurements, 197P/s)
As per the PSuU workplan, to be able to generate the measurements with an Exploratory coverage, we'll be required to be able to generate measurements across N_t=500, N_s=3 and N_sw = 2048 * 209 = 428,032, giving us a total of 2048*209*3*500 = 642,048,000 state measurements. Assuming a compute speed of 150P/s, we would require 49 days' worth of compute to generate the dataset. Having the Optimal coverage implies 3^9*2^2*50*209*5,000 = 4,113,747,000,000 state measurements, therefore requiring 317,418 days or a 6500x coverage over the Exploratory one.
Assuming a target simulation duration of two hours for the Exploratory coverage, we would require a speed-up of 588x (or ~90,000 P/s). As a hypothesis, we can assume we can have the following stacking performance multipliers by doing the following actions:
- Disabling
deepcopy:10ximprovement - Model optimizations:
2ximprovement (20xcumulative) - Simulation optimizations:
2ximprovement (40xcumulative) - Running on a cloud VM:
20ximprovement (800xcumulative)
For Optimal coverage, assuming a target simulation duration of 7 days, we would require a ~45,000x speed-up. Therefore, we may require employing heuristics or iterative approaches.