diff --git a/.azure-pipelines-templates/daily-matrix.yml b/.azure-pipelines-templates/daily-matrix.yml index 497affa2349..6e110647d70 100644 --- a/.azure-pipelines-templates/daily-matrix.yml +++ b/.azure-pipelines-templates/daily-matrix.yml @@ -5,7 +5,7 @@ parameters: pool: ado-virtual-ccf-sub SGX: container: sgx - pool: ado-sgx-ccf-sub + pool: ado-sgx-ccf-sub-backup SNPCC: container: snp pool: ado-virtual-ccf-sub diff --git a/.azure-pipelines-templates/matrix.yml b/.azure-pipelines-templates/matrix.yml index 088908e352b..6b85bc2f7d8 100644 --- a/.azure-pipelines-templates/matrix.yml +++ b/.azure-pipelines-templates/matrix.yml @@ -11,7 +11,7 @@ parameters: pool: ado-virtual-ccf-sub SGX: container: sgx - pool: ado-sgx-ccf-sub + pool: ado-sgx-ccf-sub-backup SNPCC: container: snp pool: ado-virtual-ccf-sub diff --git a/.azure-pipelines-templates/stress-matrix.yml b/.azure-pipelines-templates/stress-matrix.yml index 017ec216516..5ab5c5bf59e 100644 --- a/.azure-pipelines-templates/stress-matrix.yml +++ b/.azure-pipelines-templates/stress-matrix.yml @@ -4,7 +4,7 @@ jobs: target: SGX env: container: sgx - pool: ado-sgx-ccf-sub + pool: ado-sgx-ccf-sub-backup cmake_args: "-DCOMPILE_TARGET=sgx" suffix: "StressTest" artifact_name: "StressTest" diff --git a/tests/perf-system/submitter/submit.cpp b/tests/perf-system/submitter/submit.cpp index 4c7d3bc223e..78d1b3dc10d 100644 --- a/tests/perf-system/submitter/submit.cpp +++ b/tests/perf-system/submitter/submit.cpp @@ -18,9 +18,8 @@ #include #include #include -#include -#include #include +#include using namespace std; using namespace client; @@ -314,7 +313,7 @@ int main(int argc, char** argv) { for (size_t ridx = read_reqs; ridx < requests_size; ridx++) { - clock_gettime(CLOCK_MONOTONIC, &start[ridx]); + clock_gettime(CLOCK_REALTIME, &start[ridx]); auto request = data_handler.request[ridx]; connection->write({request.data(), request.size()}); if ( @@ -322,7 +321,7 @@ int main(int argc, char** argv) ridx - read_reqs >= args.max_inflight_requests) { responses[read_reqs] = connection->read_response(); - clock_gettime(CLOCK_MONOTONIC, &end[read_reqs]); + clock_gettime(CLOCK_REALTIME, &end[read_reqs]); read_reqs++; } if (ridx % 20000 == 0) @@ -334,7 +333,7 @@ int main(int argc, char** argv) while (read_reqs < requests_size) { responses[read_reqs] = connection->read_response(); - clock_gettime(CLOCK_MONOTONIC, &end[read_reqs]); + clock_gettime(CLOCK_REALTIME, &end[read_reqs]); read_reqs++; } connection.reset(); @@ -355,12 +354,6 @@ int main(int argc, char** argv) LOG_INFO_FMT("Finished Request Submission"); - // Calculate boot time - struct sysinfo info; - sysinfo(&info); - const auto boot_time = time(NULL) - info.uptime; - const auto boot_time_us = boot_time * 1'000'000; - for (size_t req = 0; req < requests_size; req++) { auto& response = responses[req]; @@ -378,9 +371,9 @@ int main(int argc, char** argv) data_handler.response_body.push_back(std::move(response.body)); size_t send_time_us = - boot_time_us + start[req].tv_sec * 1'000'000 + start[req].tv_nsec / 1000; + start[req].tv_sec * 1'000'000 + start[req].tv_nsec / 1000; size_t response_time_us = - boot_time_us + end[req].tv_sec * 1'000'000 + end[req].tv_nsec / 1000; + end[req].tv_sec * 1'000'000 + end[req].tv_nsec / 1000; data_handler.send_time.push_back(send_time_us); data_handler.response_time.push_back(response_time_us); }