| /* |
| * Copyright (C) 2023, Pierrick Bouvier <pierrick.bouvier@linaro.org> |
| * |
| * Demonstrates and tests usage of inline ops. |
| * |
| * License: GNU GPL, version 2 or later. |
| * See the COPYING file in the top-level directory. |
| */ |
| |
| #include <glib.h> |
| #include <stdint.h> |
| #include <stdio.h> |
| |
| #include <qemu-plugin.h> |
| |
| typedef struct { |
| uint64_t count_tb; |
| uint64_t count_tb_inline; |
| uint64_t count_insn; |
| uint64_t count_insn_inline; |
| uint64_t count_mem; |
| uint64_t count_mem_inline; |
| uint64_t tb_cond_num_trigger; |
| uint64_t tb_cond_track_count; |
| uint64_t insn_cond_num_trigger; |
| uint64_t insn_cond_track_count; |
| } CPUCount; |
| |
| static const uint64_t cond_trigger_limit = 100; |
| |
| typedef struct { |
| uint64_t data_insn; |
| uint64_t data_tb; |
| uint64_t data_mem; |
| } CPUData; |
| |
| static struct qemu_plugin_scoreboard *counts; |
| static qemu_plugin_u64 count_tb; |
| static qemu_plugin_u64 count_tb_inline; |
| static qemu_plugin_u64 count_insn; |
| static qemu_plugin_u64 count_insn_inline; |
| static qemu_plugin_u64 count_mem; |
| static qemu_plugin_u64 count_mem_inline; |
| static qemu_plugin_u64 tb_cond_num_trigger; |
| static qemu_plugin_u64 tb_cond_track_count; |
| static qemu_plugin_u64 insn_cond_num_trigger; |
| static qemu_plugin_u64 insn_cond_track_count; |
| static struct qemu_plugin_scoreboard *data; |
| static qemu_plugin_u64 data_insn; |
| static qemu_plugin_u64 data_tb; |
| static qemu_plugin_u64 data_mem; |
| |
| static uint64_t global_count_tb; |
| static uint64_t global_count_insn; |
| static uint64_t global_count_mem; |
| static unsigned int max_cpu_index; |
| static GMutex tb_lock; |
| static GMutex insn_lock; |
| static GMutex mem_lock; |
| |
| QEMU_PLUGIN_EXPORT int qemu_plugin_version = QEMU_PLUGIN_VERSION; |
| |
| static void stats_insn(void) |
| { |
| const uint64_t expected = global_count_insn; |
| const uint64_t per_vcpu = qemu_plugin_u64_sum(count_insn); |
| const uint64_t inl_per_vcpu = |
| qemu_plugin_u64_sum(count_insn_inline); |
| const uint64_t cond_num_trigger = |
| qemu_plugin_u64_sum(insn_cond_num_trigger); |
| const uint64_t cond_track_left = qemu_plugin_u64_sum(insn_cond_track_count); |
| const uint64_t conditional = |
| cond_num_trigger * cond_trigger_limit + cond_track_left; |
| g_autoptr(GString) stats = g_string_new(""); |
| g_string_append_printf(stats, "insn: %" PRIu64 "\n", expected); |
| g_string_append_printf(stats, "insn: %" PRIu64 " (per vcpu)\n", per_vcpu); |
| g_string_append_printf(stats, "insn: %" PRIu64 " (per vcpu inline)\n", inl_per_vcpu); |
| g_string_append_printf(stats, "insn: %" PRIu64 " (cond cb)\n", conditional); |
| qemu_plugin_outs(stats->str); |
| g_assert(expected > 0); |
| g_assert(per_vcpu == expected); |
| g_assert(inl_per_vcpu == expected); |
| g_assert(conditional == expected); |
| } |
| |
| static void stats_tb(void) |
| { |
| const uint64_t expected = global_count_tb; |
| const uint64_t per_vcpu = qemu_plugin_u64_sum(count_tb); |
| const uint64_t inl_per_vcpu = |
| qemu_plugin_u64_sum(count_tb_inline); |
| const uint64_t cond_num_trigger = qemu_plugin_u64_sum(tb_cond_num_trigger); |
| const uint64_t cond_track_left = qemu_plugin_u64_sum(tb_cond_track_count); |
| const uint64_t conditional = |
| cond_num_trigger * cond_trigger_limit + cond_track_left; |
| g_autoptr(GString) stats = g_string_new(""); |
| g_string_append_printf(stats, "tb: %" PRIu64 "\n", expected); |
| g_string_append_printf(stats, "tb: %" PRIu64 " (per vcpu)\n", per_vcpu); |
| g_string_append_printf(stats, "tb: %" PRIu64 " (per vcpu inline)\n", inl_per_vcpu); |
| g_string_append_printf(stats, "tb: %" PRIu64 " (conditional cb)\n", conditional); |
| qemu_plugin_outs(stats->str); |
| g_assert(expected > 0); |
| g_assert(per_vcpu == expected); |
| g_assert(inl_per_vcpu == expected); |
| g_assert(conditional == expected); |
| } |
| |
| static void stats_mem(void) |
| { |
| const uint64_t expected = global_count_mem; |
| const uint64_t per_vcpu = qemu_plugin_u64_sum(count_mem); |
| const uint64_t inl_per_vcpu = |
| qemu_plugin_u64_sum(count_mem_inline); |
| g_autoptr(GString) stats = g_string_new(""); |
| g_string_append_printf(stats, "mem: %" PRIu64 "\n", expected); |
| g_string_append_printf(stats, "mem: %" PRIu64 " (per vcpu)\n", per_vcpu); |
| g_string_append_printf(stats, "mem: %" PRIu64 " (per vcpu inline)\n", inl_per_vcpu); |
| qemu_plugin_outs(stats->str); |
| g_assert(expected > 0); |
| g_assert(per_vcpu == expected); |
| g_assert(inl_per_vcpu == expected); |
| } |
| |
| static void plugin_exit(qemu_plugin_id_t id, void *udata) |
| { |
| const unsigned int num_cpus = qemu_plugin_num_vcpus(); |
| g_autoptr(GString) stats = g_string_new(""); |
| g_assert(num_cpus == max_cpu_index + 1); |
| |
| for (int i = 0; i < num_cpus ; ++i) { |
| const uint64_t tb = qemu_plugin_u64_get(count_tb, i); |
| const uint64_t tb_inline = qemu_plugin_u64_get(count_tb_inline, i); |
| const uint64_t insn = qemu_plugin_u64_get(count_insn, i); |
| const uint64_t insn_inline = qemu_plugin_u64_get(count_insn_inline, i); |
| const uint64_t mem = qemu_plugin_u64_get(count_mem, i); |
| const uint64_t mem_inline = qemu_plugin_u64_get(count_mem_inline, i); |
| const uint64_t tb_cond_trigger = |
| qemu_plugin_u64_get(tb_cond_num_trigger, i); |
| const uint64_t tb_cond_left = |
| qemu_plugin_u64_get(tb_cond_track_count, i); |
| const uint64_t insn_cond_trigger = |
| qemu_plugin_u64_get(insn_cond_num_trigger, i); |
| const uint64_t insn_cond_left = |
| qemu_plugin_u64_get(insn_cond_track_count, i); |
| g_string_printf(stats, "cpu %d: tb (%" PRIu64 ", %" PRIu64 |
| ", %" PRIu64 " * %" PRIu64 " + %" PRIu64 |
| ") | " |
| "insn (%" PRIu64 ", %" PRIu64 |
| ", %" PRIu64 " * %" PRIu64 " + %" PRIu64 |
| ") | " |
| "mem (%" PRIu64 ", %" PRIu64 ")" |
| "\n", |
| i, |
| tb, tb_inline, |
| tb_cond_trigger, cond_trigger_limit, tb_cond_left, |
| insn, insn_inline, |
| insn_cond_trigger, cond_trigger_limit, insn_cond_left, |
| mem, mem_inline); |
| qemu_plugin_outs(stats->str); |
| g_assert(tb == tb_inline); |
| g_assert(insn == insn_inline); |
| g_assert(mem == mem_inline); |
| g_assert(tb_cond_trigger == tb / cond_trigger_limit); |
| g_assert(tb_cond_left == tb % cond_trigger_limit); |
| g_assert(insn_cond_trigger == insn / cond_trigger_limit); |
| g_assert(insn_cond_left == insn % cond_trigger_limit); |
| } |
| |
| stats_tb(); |
| stats_insn(); |
| stats_mem(); |
| |
| qemu_plugin_scoreboard_free(counts); |
| qemu_plugin_scoreboard_free(data); |
| } |
| |
| static void vcpu_tb_exec(unsigned int cpu_index, void *udata) |
| { |
| qemu_plugin_u64_add(count_tb, cpu_index, 1); |
| g_assert(qemu_plugin_u64_get(data_tb, cpu_index) == (uintptr_t) udata); |
| g_mutex_lock(&tb_lock); |
| max_cpu_index = MAX(max_cpu_index, cpu_index); |
| global_count_tb++; |
| g_mutex_unlock(&tb_lock); |
| } |
| |
| static void vcpu_tb_cond_exec(unsigned int cpu_index, void *udata) |
| { |
| g_assert(qemu_plugin_u64_get(tb_cond_track_count, cpu_index) == |
| cond_trigger_limit); |
| g_assert(qemu_plugin_u64_get(data_tb, cpu_index) == (uintptr_t) udata); |
| qemu_plugin_u64_set(tb_cond_track_count, cpu_index, 0); |
| qemu_plugin_u64_add(tb_cond_num_trigger, cpu_index, 1); |
| } |
| |
| static void vcpu_insn_cond_exec(unsigned int cpu_index, void *udata) |
| { |
| g_assert(qemu_plugin_u64_get(insn_cond_track_count, cpu_index) == |
| cond_trigger_limit); |
| g_assert(qemu_plugin_u64_get(data_insn, cpu_index) == (uintptr_t) udata); |
| qemu_plugin_u64_set(insn_cond_track_count, cpu_index, 0); |
| qemu_plugin_u64_add(insn_cond_num_trigger, cpu_index, 1); |
| } |
| |
| static void vcpu_insn_exec(unsigned int cpu_index, void *udata) |
| { |
| qemu_plugin_u64_add(count_insn, cpu_index, 1); |
| g_assert(qemu_plugin_u64_get(data_insn, cpu_index) == (uintptr_t) udata); |
| g_mutex_lock(&insn_lock); |
| global_count_insn++; |
| g_mutex_unlock(&insn_lock); |
| } |
| |
| static void vcpu_mem_access(unsigned int cpu_index, |
| qemu_plugin_meminfo_t info, |
| uint64_t vaddr, |
| void *udata) |
| { |
| qemu_plugin_u64_add(count_mem, cpu_index, 1); |
| g_assert(qemu_plugin_u64_get(data_mem, cpu_index) == (uintptr_t) udata); |
| g_mutex_lock(&mem_lock); |
| global_count_mem++; |
| g_mutex_unlock(&mem_lock); |
| } |
| |
| static void vcpu_tb_trans(qemu_plugin_id_t id, struct qemu_plugin_tb *tb) |
| { |
| void *tb_store = tb; |
| qemu_plugin_register_vcpu_tb_exec_inline_per_vcpu( |
| tb, QEMU_PLUGIN_INLINE_STORE_U64, data_tb, (uintptr_t) tb_store); |
| qemu_plugin_register_vcpu_tb_exec_cb( |
| tb, vcpu_tb_exec, QEMU_PLUGIN_CB_NO_REGS, tb_store); |
| qemu_plugin_register_vcpu_tb_exec_inline_per_vcpu( |
| tb, QEMU_PLUGIN_INLINE_ADD_U64, count_tb_inline, 1); |
| |
| qemu_plugin_register_vcpu_tb_exec_inline_per_vcpu( |
| tb, QEMU_PLUGIN_INLINE_ADD_U64, tb_cond_track_count, 1); |
| qemu_plugin_register_vcpu_tb_exec_cond_cb( |
| tb, vcpu_tb_cond_exec, QEMU_PLUGIN_CB_NO_REGS, |
| QEMU_PLUGIN_COND_EQ, tb_cond_track_count, cond_trigger_limit, tb_store); |
| |
| for (int idx = 0; idx < qemu_plugin_tb_n_insns(tb); ++idx) { |
| struct qemu_plugin_insn *insn = qemu_plugin_tb_get_insn(tb, idx); |
| void *insn_store = insn; |
| void *mem_store = (char *)insn_store + 0xff; |
| |
| qemu_plugin_register_vcpu_insn_exec_inline_per_vcpu( |
| insn, QEMU_PLUGIN_INLINE_STORE_U64, data_insn, |
| (uintptr_t) insn_store); |
| qemu_plugin_register_vcpu_insn_exec_cb( |
| insn, vcpu_insn_exec, QEMU_PLUGIN_CB_NO_REGS, insn_store); |
| qemu_plugin_register_vcpu_insn_exec_inline_per_vcpu( |
| insn, QEMU_PLUGIN_INLINE_ADD_U64, count_insn_inline, 1); |
| |
| qemu_plugin_register_vcpu_insn_exec_inline_per_vcpu( |
| insn, QEMU_PLUGIN_INLINE_ADD_U64, insn_cond_track_count, 1); |
| qemu_plugin_register_vcpu_insn_exec_cond_cb( |
| insn, vcpu_insn_cond_exec, QEMU_PLUGIN_CB_NO_REGS, |
| QEMU_PLUGIN_COND_EQ, insn_cond_track_count, cond_trigger_limit, |
| insn_store); |
| |
| qemu_plugin_register_vcpu_mem_inline_per_vcpu( |
| insn, QEMU_PLUGIN_MEM_RW, |
| QEMU_PLUGIN_INLINE_STORE_U64, |
| data_mem, (uintptr_t) mem_store); |
| qemu_plugin_register_vcpu_mem_cb(insn, &vcpu_mem_access, |
| QEMU_PLUGIN_CB_NO_REGS, |
| QEMU_PLUGIN_MEM_RW, mem_store); |
| qemu_plugin_register_vcpu_mem_inline_per_vcpu( |
| insn, QEMU_PLUGIN_MEM_RW, |
| QEMU_PLUGIN_INLINE_ADD_U64, |
| count_mem_inline, 1); |
| } |
| } |
| |
| QEMU_PLUGIN_EXPORT |
| int qemu_plugin_install(qemu_plugin_id_t id, const qemu_info_t *info, |
| int argc, char **argv) |
| { |
| counts = qemu_plugin_scoreboard_new(sizeof(CPUCount)); |
| count_tb = qemu_plugin_scoreboard_u64_in_struct( |
| counts, CPUCount, count_tb); |
| count_insn = qemu_plugin_scoreboard_u64_in_struct( |
| counts, CPUCount, count_insn); |
| count_mem = qemu_plugin_scoreboard_u64_in_struct( |
| counts, CPUCount, count_mem); |
| count_tb_inline = qemu_plugin_scoreboard_u64_in_struct( |
| counts, CPUCount, count_tb_inline); |
| count_insn_inline = qemu_plugin_scoreboard_u64_in_struct( |
| counts, CPUCount, count_insn_inline); |
| count_mem_inline = qemu_plugin_scoreboard_u64_in_struct( |
| counts, CPUCount, count_mem_inline); |
| tb_cond_num_trigger = qemu_plugin_scoreboard_u64_in_struct( |
| counts, CPUCount, tb_cond_num_trigger); |
| tb_cond_track_count = qemu_plugin_scoreboard_u64_in_struct( |
| counts, CPUCount, tb_cond_track_count); |
| insn_cond_num_trigger = qemu_plugin_scoreboard_u64_in_struct( |
| counts, CPUCount, insn_cond_num_trigger); |
| insn_cond_track_count = qemu_plugin_scoreboard_u64_in_struct( |
| counts, CPUCount, insn_cond_track_count); |
| data = qemu_plugin_scoreboard_new(sizeof(CPUData)); |
| data_insn = qemu_plugin_scoreboard_u64_in_struct(data, CPUData, data_insn); |
| data_tb = qemu_plugin_scoreboard_u64_in_struct(data, CPUData, data_tb); |
| data_mem = qemu_plugin_scoreboard_u64_in_struct(data, CPUData, data_mem); |
| |
| qemu_plugin_register_vcpu_tb_trans_cb(id, vcpu_tb_trans); |
| qemu_plugin_register_atexit_cb(id, plugin_exit, NULL); |
| |
| return 0; |
| } |