| /* |
| * SPDX-License-Identifier: BSD-2-Clause |
| * |
| * Copyright (c) 2021 Western Digital Corporation or its affiliates. |
| * |
| * Authors: |
| * Atish Patra <atish.patra@wdc.com> |
| */ |
| |
| #include <sbi/riscv_asm.h> |
| #include <sbi/sbi_bitops.h> |
| #include <sbi/sbi_console.h> |
| #include <sbi/sbi_ecall_interface.h> |
| #include <sbi/sbi_hart.h> |
| #include <sbi/sbi_heap.h> |
| #include <sbi/sbi_platform.h> |
| #include <sbi/sbi_pmu.h> |
| #include <sbi/sbi_scratch.h> |
| #include <sbi/sbi_string.h> |
| |
| /** Information about hardware counters */ |
| struct sbi_pmu_hw_event { |
| uint32_t counters; |
| uint32_t start_idx; |
| uint32_t end_idx; |
| /* Event selector value used only for raw events. The event select value |
| * can be a even id or a selector value for set of events encoded in few |
| * bits. In case latter, the bits used for encoding of the events should |
| * be zeroed out in the select value. |
| */ |
| uint64_t select; |
| /** |
| * The select_mask indicates which bits are encoded for the event(s). |
| */ |
| uint64_t select_mask; |
| }; |
| |
| /* Information about PMU counters as per SBI specification */ |
| union sbi_pmu_ctr_info { |
| unsigned long value; |
| struct { |
| unsigned long csr:12; |
| unsigned long width:6; |
| #if __riscv_xlen == 32 |
| unsigned long reserved:13; |
| #else |
| unsigned long reserved:45; |
| #endif |
| unsigned long type:1; |
| }; |
| }; |
| |
| #if SBI_PMU_FW_CTR_MAX >= BITS_PER_LONG |
| #error "Can't handle firmware counters beyond BITS_PER_LONG" |
| #endif |
| |
| /** Per-HART state of the PMU counters */ |
| struct sbi_pmu_hart_state { |
| /* HART to which this state belongs */ |
| uint32_t hartid; |
| /* Counter to enabled event mapping */ |
| uint32_t active_events[SBI_PMU_HW_CTR_MAX + SBI_PMU_FW_CTR_MAX]; |
| /* Bitmap of firmware counters started */ |
| unsigned long fw_counters_started; |
| /* |
| * Counter values for SBI firmware events and event codes |
| * for platform firmware events. Both are mutually exclusive |
| * and hence can optimally share the same memory. |
| */ |
| uint64_t fw_counters_data[SBI_PMU_FW_CTR_MAX]; |
| }; |
| |
| /** Offset of pointer to PMU HART state in scratch space */ |
| static unsigned long phs_ptr_offset; |
| |
| #define pmu_get_hart_state_ptr(__scratch) \ |
| sbi_scratch_read_type((__scratch), void *, phs_ptr_offset) |
| |
| #define pmu_thishart_state_ptr() \ |
| pmu_get_hart_state_ptr(sbi_scratch_thishart_ptr()) |
| |
| #define pmu_set_hart_state_ptr(__scratch, __phs) \ |
| sbi_scratch_write_type((__scratch), void *, phs_ptr_offset, (__phs)) |
| |
| /* Platform specific PMU device */ |
| static const struct sbi_pmu_device *pmu_dev = NULL; |
| |
| /* Mapping between event range and possible counters */ |
| static struct sbi_pmu_hw_event *hw_event_map; |
| |
| /* Maximum number of hardware events available */ |
| static uint32_t num_hw_events; |
| /* Maximum number of hardware counters available */ |
| static uint32_t num_hw_ctrs; |
| |
| /* Maximum number of counters available */ |
| static uint32_t total_ctrs; |
| |
| /* Helper macros to retrieve event idx and code type */ |
| #define get_cidx_type(x) \ |
| (((x) & SBI_PMU_EVENT_IDX_TYPE_MASK) >> SBI_PMU_EVENT_IDX_TYPE_OFFSET) |
| #define get_cidx_code(x) (x & SBI_PMU_EVENT_IDX_CODE_MASK) |
| |
| /** |
| * Perform a sanity check on event & counter mappings with event range overlap check |
| * @param evtA Pointer to the existing hw event structure |
| * @param evtB Pointer to the new hw event structure |
| * |
| * Return false if the range doesn't overlap, true otherwise |
| */ |
| static bool pmu_event_range_overlap(struct sbi_pmu_hw_event *evtA, |
| struct sbi_pmu_hw_event *evtB) |
| { |
| /* check if the range of events overlap with a previous entry */ |
| if (((evtA->end_idx < evtB->start_idx) && (evtA->end_idx < evtB->end_idx)) || |
| ((evtA->start_idx > evtB->start_idx) && (evtA->start_idx > evtB->end_idx))) |
| return false; |
| return true; |
| } |
| |
| static bool pmu_event_select_overlap(struct sbi_pmu_hw_event *evt, |
| uint64_t select_val, uint64_t select_mask) |
| { |
| if ((evt->select == select_val) && (evt->select_mask == select_mask)) |
| return true; |
| |
| return false; |
| } |
| |
| static int pmu_event_validate(struct sbi_pmu_hart_state *phs, |
| unsigned long event_idx, uint64_t edata) |
| { |
| uint32_t event_idx_type = get_cidx_type(event_idx); |
| uint32_t event_idx_code = get_cidx_code(event_idx); |
| uint32_t event_idx_code_max = -1; |
| uint32_t cache_ops_result, cache_ops_id, cache_id; |
| |
| switch(event_idx_type) { |
| case SBI_PMU_EVENT_TYPE_HW: |
| event_idx_code_max = SBI_PMU_HW_GENERAL_MAX; |
| break; |
| case SBI_PMU_EVENT_TYPE_FW: |
| if ((event_idx_code >= SBI_PMU_FW_MAX && |
| event_idx_code <= SBI_PMU_FW_RESERVED_MAX) || |
| event_idx_code > SBI_PMU_FW_PLATFORM) |
| return SBI_EINVAL; |
| |
| if (SBI_PMU_FW_PLATFORM == event_idx_code && |
| pmu_dev && pmu_dev->fw_event_validate_encoding) |
| return pmu_dev->fw_event_validate_encoding(phs->hartid, |
| edata); |
| else |
| event_idx_code_max = SBI_PMU_FW_MAX; |
| break; |
| case SBI_PMU_EVENT_TYPE_HW_CACHE: |
| cache_ops_result = event_idx_code & |
| SBI_PMU_EVENT_HW_CACHE_OPS_RESULT; |
| cache_ops_id = (event_idx_code & |
| SBI_PMU_EVENT_HW_CACHE_OPS_ID_MASK) >> |
| SBI_PMU_EVENT_HW_CACHE_OPS_ID_OFFSET; |
| cache_id = (event_idx_code & |
| SBI_PMU_EVENT_HW_CACHE_ID_MASK) >> |
| SBI_PMU_EVENT_HW_CACHE_ID_OFFSET; |
| if ((cache_ops_result < SBI_PMU_HW_CACHE_RESULT_MAX) && |
| (cache_ops_id < SBI_PMU_HW_CACHE_OP_MAX) && |
| (cache_id < SBI_PMU_HW_CACHE_MAX)) |
| return event_idx_type; |
| else |
| return SBI_EINVAL; |
| break; |
| case SBI_PMU_EVENT_TYPE_HW_RAW: |
| event_idx_code_max = 1; // event_idx.code should be zero |
| break; |
| default: |
| return SBI_EINVAL; |
| } |
| |
| if (event_idx_code < event_idx_code_max) |
| return event_idx_type; |
| |
| return SBI_EINVAL; |
| } |
| |
| static int pmu_ctr_validate(struct sbi_pmu_hart_state *phs, |
| uint32_t cidx, uint32_t *event_idx_code) |
| { |
| uint32_t event_idx_val; |
| uint32_t event_idx_type; |
| |
| if (cidx >= total_ctrs) |
| return SBI_EINVAL; |
| |
| event_idx_val = phs->active_events[cidx]; |
| event_idx_type = get_cidx_type(event_idx_val); |
| if (event_idx_val == SBI_PMU_EVENT_IDX_INVALID || |
| event_idx_type >= SBI_PMU_EVENT_TYPE_MAX) |
| return SBI_EINVAL; |
| |
| *event_idx_code = get_cidx_code(event_idx_val); |
| |
| return event_idx_type; |
| } |
| |
| int sbi_pmu_ctr_fw_read(uint32_t cidx, uint64_t *cval) |
| { |
| int event_idx_type; |
| uint32_t event_code; |
| struct sbi_pmu_hart_state *phs = pmu_thishart_state_ptr(); |
| |
| event_idx_type = pmu_ctr_validate(phs, cidx, &event_code); |
| if (event_idx_type != SBI_PMU_EVENT_TYPE_FW) |
| return SBI_EINVAL; |
| |
| if ((event_code >= SBI_PMU_FW_MAX && |
| event_code <= SBI_PMU_FW_RESERVED_MAX) || |
| event_code > SBI_PMU_FW_PLATFORM) |
| return SBI_EINVAL; |
| |
| if (SBI_PMU_FW_PLATFORM == event_code) { |
| if (pmu_dev && pmu_dev->fw_counter_read_value) |
| *cval = pmu_dev->fw_counter_read_value(phs->hartid, |
| cidx - |
| num_hw_ctrs); |
| else |
| *cval = 0; |
| } else |
| *cval = phs->fw_counters_data[cidx - num_hw_ctrs]; |
| |
| return 0; |
| } |
| |
| static int pmu_add_hw_event_map(u32 eidx_start, u32 eidx_end, u32 cmap, |
| uint64_t select, uint64_t select_mask) |
| { |
| int i = 0; |
| bool is_overlap; |
| struct sbi_pmu_hw_event *event = &hw_event_map[num_hw_events]; |
| struct sbi_scratch *scratch = sbi_scratch_thishart_ptr(); |
| uint32_t ctr_avail_mask = sbi_hart_mhpm_mask(scratch) | 0x7; |
| |
| /* The first two counters are reserved by priv spec */ |
| if (eidx_start > SBI_PMU_HW_INSTRUCTIONS && (cmap & SBI_PMU_FIXED_CTR_MASK)) |
| return SBI_EDENIED; |
| |
| if (num_hw_events >= SBI_PMU_HW_EVENT_MAX - 1) { |
| sbi_printf("Can not handle more than %d perf events\n", |
| SBI_PMU_HW_EVENT_MAX); |
| return SBI_EFAIL; |
| } |
| |
| event->start_idx = eidx_start; |
| event->end_idx = eidx_end; |
| |
| /* Sanity check */ |
| for (i = 0; i < num_hw_events; i++) { |
| if (eidx_start == SBI_PMU_EVENT_RAW_IDX) |
| /* All raw events have same event idx. Just do sanity check on select */ |
| is_overlap = pmu_event_select_overlap(&hw_event_map[i], |
| select, select_mask); |
| else |
| is_overlap = pmu_event_range_overlap(&hw_event_map[i], event); |
| if (is_overlap) |
| goto reset_event; |
| } |
| |
| event->select_mask = select_mask; |
| /* Map the only the counters that are available in the hardware */ |
| event->counters = cmap & ctr_avail_mask; |
| event->select = select; |
| num_hw_events++; |
| |
| return 0; |
| |
| reset_event: |
| event->start_idx = 0; |
| event->end_idx = 0; |
| return SBI_EINVAL; |
| } |
| |
| /** |
| * Logical counter ids are assigned to hardware counters are assigned consecutively. |
| * E.g. counter0 must count MCYCLE where counter2 must count minstret. Similarly, |
| * counterX will mhpmcounterX. |
| */ |
| int sbi_pmu_add_hw_event_counter_map(u32 eidx_start, u32 eidx_end, u32 cmap) |
| { |
| if ((eidx_start > eidx_end) || eidx_start == SBI_PMU_EVENT_RAW_IDX || |
| eidx_end == SBI_PMU_EVENT_RAW_IDX) |
| return SBI_EINVAL; |
| |
| return pmu_add_hw_event_map(eidx_start, eidx_end, cmap, 0, 0); |
| } |
| |
| int sbi_pmu_add_raw_event_counter_map(uint64_t select, uint64_t select_mask, u32 cmap) |
| { |
| return pmu_add_hw_event_map(SBI_PMU_EVENT_RAW_IDX, |
| SBI_PMU_EVENT_RAW_IDX, cmap, select, select_mask); |
| } |
| |
| static int pmu_ctr_enable_irq_hw(int ctr_idx) |
| { |
| unsigned long mhpmevent_csr; |
| unsigned long mhpmevent_curr; |
| unsigned long mip_val; |
| unsigned long of_mask; |
| |
| if (ctr_idx < 3 || ctr_idx >= SBI_PMU_HW_CTR_MAX) |
| return SBI_EFAIL; |
| |
| #if __riscv_xlen == 32 |
| mhpmevent_csr = CSR_MHPMEVENT3H + ctr_idx - 3; |
| of_mask = (uint32_t)~MHPMEVENTH_OF; |
| #else |
| mhpmevent_csr = CSR_MHPMEVENT3 + ctr_idx - 3; |
| of_mask = ~MHPMEVENT_OF; |
| #endif |
| |
| mhpmevent_curr = csr_read_num(mhpmevent_csr); |
| mip_val = csr_read(CSR_MIP); |
| /** |
| * Clear out the OF bit so that next interrupt can be enabled. |
| * This should be done only when the corresponding overflow interrupt |
| * bit is cleared. That indicates that software has already handled the |
| * previous interrupts or the hardware yet to set an overflow interrupt. |
| * Otherwise, there will be race conditions where we may clear the bit |
| * the software is yet to handle the interrupt. |
| */ |
| if (!(mip_val & MIP_LCOFIP)) { |
| mhpmevent_curr &= of_mask; |
| csr_write_num(mhpmevent_csr, mhpmevent_curr); |
| } |
| |
| return 0; |
| } |
| |
| static void pmu_ctr_write_hw(uint32_t cidx, uint64_t ival) |
| { |
| #if __riscv_xlen == 32 |
| csr_write_num(CSR_MCYCLE + cidx, 0); |
| csr_write_num(CSR_MCYCLE + cidx, ival & 0xFFFFFFFF); |
| csr_write_num(CSR_MCYCLEH + cidx, ival >> BITS_PER_LONG); |
| #else |
| csr_write_num(CSR_MCYCLE + cidx, ival); |
| #endif |
| } |
| |
| static int pmu_ctr_start_hw(uint32_t cidx, uint64_t ival, bool ival_update) |
| { |
| struct sbi_scratch *scratch = sbi_scratch_thishart_ptr(); |
| unsigned long mctr_inhbt; |
| |
| /* Make sure the counter index lies within the range and is not TM bit */ |
| if (cidx >= num_hw_ctrs || cidx == 1) |
| return SBI_EINVAL; |
| |
| if (sbi_hart_priv_version(scratch) < SBI_HART_PRIV_VER_1_11) { |
| if (ival_update) |
| pmu_ctr_write_hw(cidx, ival); |
| return 0; |
| } |
| |
| /* |
| * Some of the hardware may not support mcountinhibit but perf stat |
| * still can work if supervisor mode programs the initial value. |
| */ |
| mctr_inhbt = csr_read(CSR_MCOUNTINHIBIT); |
| if (!__test_bit(cidx, &mctr_inhbt)) |
| return SBI_EALREADY_STARTED; |
| |
| __clear_bit(cidx, &mctr_inhbt); |
| |
| if (sbi_hart_has_extension(scratch, SBI_HART_EXT_SSCOFPMF)) |
| pmu_ctr_enable_irq_hw(cidx); |
| if (ival_update) |
| pmu_ctr_write_hw(cidx, ival); |
| if (pmu_dev && pmu_dev->hw_counter_enable_irq) |
| pmu_dev->hw_counter_enable_irq(cidx); |
| |
| csr_write(CSR_MCOUNTINHIBIT, mctr_inhbt); |
| |
| return 0; |
| } |
| |
| int sbi_pmu_irq_bit(void) |
| { |
| struct sbi_scratch *scratch = sbi_scratch_thishart_ptr(); |
| |
| if (sbi_hart_has_extension(scratch, SBI_HART_EXT_SSCOFPMF)) |
| return MIP_LCOFIP; |
| if (pmu_dev && pmu_dev->hw_counter_irq_bit) |
| return pmu_dev->hw_counter_irq_bit(); |
| |
| return 0; |
| } |
| |
| static int pmu_ctr_start_fw(struct sbi_pmu_hart_state *phs, |
| uint32_t cidx, uint32_t event_code, |
| uint64_t event_data, uint64_t ival, |
| bool ival_update) |
| { |
| if ((event_code >= SBI_PMU_FW_MAX && |
| event_code <= SBI_PMU_FW_RESERVED_MAX) || |
| event_code > SBI_PMU_FW_PLATFORM) |
| return SBI_EINVAL; |
| |
| if (SBI_PMU_FW_PLATFORM == event_code) { |
| if (!pmu_dev || |
| !pmu_dev->fw_counter_write_value || |
| !pmu_dev->fw_counter_start) { |
| return SBI_EINVAL; |
| } |
| |
| if (ival_update) |
| pmu_dev->fw_counter_write_value(phs->hartid, |
| cidx - num_hw_ctrs, |
| ival); |
| |
| return pmu_dev->fw_counter_start(phs->hartid, |
| cidx - num_hw_ctrs, |
| event_data); |
| } else { |
| if (ival_update) |
| phs->fw_counters_data[cidx - num_hw_ctrs] = ival; |
| } |
| |
| phs->fw_counters_started |= BIT(cidx - num_hw_ctrs); |
| |
| return 0; |
| } |
| |
| int sbi_pmu_ctr_start(unsigned long cbase, unsigned long cmask, |
| unsigned long flags, uint64_t ival) |
| { |
| struct sbi_pmu_hart_state *phs = pmu_thishart_state_ptr(); |
| int event_idx_type; |
| uint32_t event_code; |
| int ret = SBI_EINVAL; |
| bool bUpdate = false; |
| int i, cidx; |
| uint64_t edata; |
| |
| if ((cbase + sbi_fls(cmask)) >= total_ctrs) |
| return ret; |
| |
| if (flags & SBI_PMU_START_FLAG_SET_INIT_VALUE) |
| bUpdate = true; |
| |
| for_each_set_bit(i, &cmask, BITS_PER_LONG) { |
| cidx = i + cbase; |
| event_idx_type = pmu_ctr_validate(phs, cidx, &event_code); |
| if (event_idx_type < 0) |
| /* Continue the start operation for other counters */ |
| continue; |
| else if (event_idx_type == SBI_PMU_EVENT_TYPE_FW) { |
| edata = (event_code == SBI_PMU_FW_PLATFORM) ? |
| phs->fw_counters_data[cidx - num_hw_ctrs] |
| : 0x0; |
| ret = pmu_ctr_start_fw(phs, cidx, event_code, edata, |
| ival, bUpdate); |
| } |
| else |
| ret = pmu_ctr_start_hw(cidx, ival, bUpdate); |
| } |
| |
| return ret; |
| } |
| |
| static int pmu_ctr_stop_hw(uint32_t cidx) |
| { |
| struct sbi_scratch *scratch = sbi_scratch_thishart_ptr(); |
| unsigned long mctr_inhbt; |
| |
| if (sbi_hart_priv_version(scratch) < SBI_HART_PRIV_VER_1_11) |
| return 0; |
| |
| mctr_inhbt = csr_read(CSR_MCOUNTINHIBIT); |
| |
| /* Make sure the counter index lies within the range and is not TM bit */ |
| if (cidx >= num_hw_ctrs || cidx == 1) |
| return SBI_EINVAL; |
| |
| if (!__test_bit(cidx, &mctr_inhbt)) { |
| __set_bit(cidx, &mctr_inhbt); |
| csr_write(CSR_MCOUNTINHIBIT, mctr_inhbt); |
| if (pmu_dev && pmu_dev->hw_counter_disable_irq) { |
| pmu_dev->hw_counter_disable_irq(cidx); |
| } |
| return 0; |
| } else |
| return SBI_EALREADY_STOPPED; |
| } |
| |
| static int pmu_ctr_stop_fw(struct sbi_pmu_hart_state *phs, |
| uint32_t cidx, uint32_t event_code) |
| { |
| int ret; |
| |
| if ((event_code >= SBI_PMU_FW_MAX && |
| event_code <= SBI_PMU_FW_RESERVED_MAX) || |
| event_code > SBI_PMU_FW_PLATFORM) |
| return SBI_EINVAL; |
| |
| if (SBI_PMU_FW_PLATFORM == event_code && |
| pmu_dev && pmu_dev->fw_counter_stop) { |
| ret = pmu_dev->fw_counter_stop(phs->hartid, cidx - num_hw_ctrs); |
| if (ret) |
| return ret; |
| } |
| |
| phs->fw_counters_started &= ~BIT(cidx - num_hw_ctrs); |
| |
| return 0; |
| } |
| |
| static int pmu_reset_hw_mhpmevent(int ctr_idx) |
| { |
| if (ctr_idx < 3 || ctr_idx >= SBI_PMU_HW_CTR_MAX) |
| return SBI_EFAIL; |
| #if __riscv_xlen == 32 |
| csr_write_num(CSR_MHPMEVENT3 + ctr_idx - 3, 0); |
| if (sbi_hart_has_extension(sbi_scratch_thishart_ptr(), |
| SBI_HART_EXT_SSCOFPMF)) |
| csr_write_num(CSR_MHPMEVENT3H + ctr_idx - 3, 0); |
| #else |
| csr_write_num(CSR_MHPMEVENT3 + ctr_idx - 3, 0); |
| #endif |
| |
| return 0; |
| } |
| |
| int sbi_pmu_ctr_stop(unsigned long cbase, unsigned long cmask, |
| unsigned long flag) |
| { |
| struct sbi_pmu_hart_state *phs = pmu_thishart_state_ptr(); |
| int ret = SBI_EINVAL; |
| int event_idx_type; |
| uint32_t event_code; |
| int i, cidx; |
| |
| if ((cbase + sbi_fls(cmask)) >= total_ctrs) |
| return SBI_EINVAL; |
| |
| for_each_set_bit(i, &cmask, BITS_PER_LONG) { |
| cidx = i + cbase; |
| event_idx_type = pmu_ctr_validate(phs, cidx, &event_code); |
| if (event_idx_type < 0) |
| /* Continue the stop operation for other counters */ |
| continue; |
| |
| else if (event_idx_type == SBI_PMU_EVENT_TYPE_FW) |
| ret = pmu_ctr_stop_fw(phs, cidx, event_code); |
| else |
| ret = pmu_ctr_stop_hw(cidx); |
| |
| if (cidx > (CSR_INSTRET - CSR_CYCLE) && flag & SBI_PMU_STOP_FLAG_RESET) { |
| phs->active_events[cidx] = SBI_PMU_EVENT_IDX_INVALID; |
| pmu_reset_hw_mhpmevent(cidx); |
| } |
| } |
| |
| return ret; |
| } |
| |
| static void pmu_update_inhibit_flags(unsigned long flags, uint64_t *mhpmevent_val) |
| { |
| if (flags & SBI_PMU_CFG_FLAG_SET_VUINH) |
| *mhpmevent_val |= MHPMEVENT_VUINH; |
| if (flags & SBI_PMU_CFG_FLAG_SET_VSINH) |
| *mhpmevent_val |= MHPMEVENT_VSINH; |
| if (flags & SBI_PMU_CFG_FLAG_SET_UINH) |
| *mhpmevent_val |= MHPMEVENT_UINH; |
| if (flags & SBI_PMU_CFG_FLAG_SET_SINH) |
| *mhpmevent_val |= MHPMEVENT_SINH; |
| } |
| |
| static int pmu_update_hw_mhpmevent(struct sbi_pmu_hw_event *hw_evt, int ctr_idx, |
| unsigned long flags, unsigned long eindex, |
| uint64_t data) |
| { |
| struct sbi_scratch *scratch = sbi_scratch_thishart_ptr(); |
| const struct sbi_platform *plat = sbi_platform_ptr(scratch); |
| uint64_t mhpmevent_val; |
| |
| /* Get the final mhpmevent value to be written from platform */ |
| mhpmevent_val = sbi_platform_pmu_xlate_to_mhpmevent(plat, eindex, data); |
| |
| if (!mhpmevent_val || ctr_idx < 3 || ctr_idx >= SBI_PMU_HW_CTR_MAX) |
| return SBI_EFAIL; |
| |
| /** |
| * Always set the OVF bit(disable interrupts) and inhibit counting of |
| * events in M-mode. The OVF bit should be enabled during the start call. |
| */ |
| if (sbi_hart_has_extension(scratch, SBI_HART_EXT_SSCOFPMF)) |
| mhpmevent_val = (mhpmevent_val & ~MHPMEVENT_SSCOF_MASK) | |
| MHPMEVENT_MINH | MHPMEVENT_OF; |
| |
| if (pmu_dev && pmu_dev->hw_counter_disable_irq) |
| pmu_dev->hw_counter_disable_irq(ctr_idx); |
| |
| /* Update the inhibit flags based on inhibit flags received from supervisor */ |
| pmu_update_inhibit_flags(flags, &mhpmevent_val); |
| |
| #if __riscv_xlen == 32 |
| csr_write_num(CSR_MHPMEVENT3 + ctr_idx - 3, mhpmevent_val & 0xFFFFFFFF); |
| if (sbi_hart_has_extension(scratch, SBI_HART_EXT_SSCOFPMF)) |
| csr_write_num(CSR_MHPMEVENT3H + ctr_idx - 3, |
| mhpmevent_val >> BITS_PER_LONG); |
| #else |
| csr_write_num(CSR_MHPMEVENT3 + ctr_idx - 3, mhpmevent_val); |
| #endif |
| |
| return 0; |
| } |
| |
| static int pmu_fixed_ctr_update_inhibit_bits(int fixed_ctr, unsigned long flags) |
| { |
| struct sbi_scratch *scratch = sbi_scratch_thishart_ptr(); |
| uint64_t cfg_val = 0, cfg_csr_no; |
| #if __riscv_xlen == 32 |
| uint64_t cfgh_csr_no; |
| #endif |
| if (!sbi_hart_has_extension(scratch, SBI_HART_EXT_SMCNTRPMF)) |
| return fixed_ctr; |
| |
| switch (fixed_ctr) { |
| case 0: |
| cfg_csr_no = CSR_MCYCLECFG; |
| #if __riscv_xlen == 32 |
| cfgh_csr_no = CSR_MCYCLECFGH; |
| #endif |
| break; |
| case 2: |
| cfg_csr_no = CSR_MINSTRETCFG; |
| #if __riscv_xlen == 32 |
| cfgh_csr_no = CSR_MINSTRETCFGH; |
| #endif |
| break; |
| default: |
| return SBI_EFAIL; |
| } |
| |
| cfg_val |= MHPMEVENT_MINH; |
| pmu_update_inhibit_flags(flags, &cfg_val); |
| #if __riscv_xlen == 32 |
| csr_write_num(cfg_csr_no, cfg_val & 0xFFFFFFFF); |
| csr_write_num(cfgh_csr_no, cfg_val >> BITS_PER_LONG); |
| #else |
| csr_write_num(cfg_csr_no, cfg_val); |
| #endif |
| return fixed_ctr; |
| } |
| |
| static int pmu_ctr_find_fixed_hw(unsigned long evt_idx_code) |
| { |
| /* Non-programmables counters are enabled always. No need to do lookup */ |
| if (evt_idx_code == SBI_PMU_HW_CPU_CYCLES) |
| return 0; |
| else if (evt_idx_code == SBI_PMU_HW_INSTRUCTIONS) |
| return 2; |
| else |
| return SBI_EINVAL; |
| } |
| |
| static int pmu_ctr_find_hw(struct sbi_pmu_hart_state *phs, |
| unsigned long cbase, unsigned long cmask, |
| unsigned long flags, |
| unsigned long event_idx, uint64_t data) |
| { |
| unsigned long ctr_mask; |
| int i, ret = 0, fixed_ctr, ctr_idx = SBI_ENOTSUPP; |
| struct sbi_pmu_hw_event *temp; |
| unsigned long mctr_inhbt = 0; |
| struct sbi_scratch *scratch = sbi_scratch_thishart_ptr(); |
| |
| if (cbase >= num_hw_ctrs) |
| return SBI_EINVAL; |
| |
| /** |
| * If Sscof is present try to find the programmable counter for |
| * cycle/instret as well. |
| */ |
| fixed_ctr = pmu_ctr_find_fixed_hw(event_idx); |
| if (fixed_ctr >= 0 && |
| !sbi_hart_has_extension(scratch, SBI_HART_EXT_SSCOFPMF)) |
| return pmu_fixed_ctr_update_inhibit_bits(fixed_ctr, flags); |
| |
| if (sbi_hart_priv_version(scratch) >= SBI_HART_PRIV_VER_1_11) |
| mctr_inhbt = csr_read(CSR_MCOUNTINHIBIT); |
| for (i = 0; i < num_hw_events; i++) { |
| temp = &hw_event_map[i]; |
| if ((temp->start_idx > event_idx && event_idx < temp->end_idx) || |
| (temp->start_idx < event_idx && event_idx > temp->end_idx)) |
| continue; |
| |
| /* For raw events, event data is used as the select value */ |
| if (event_idx == SBI_PMU_EVENT_RAW_IDX) { |
| uint64_t select_mask = temp->select_mask; |
| |
| /* The non-event map bits of data should match the selector */ |
| if (temp->select != (data & select_mask)) |
| continue; |
| } |
| /* Fixed counters should not be part of the search */ |
| ctr_mask = temp->counters & (cmask << cbase) & |
| (~SBI_PMU_FIXED_CTR_MASK); |
| for_each_set_bit_from(cbase, &ctr_mask, SBI_PMU_HW_CTR_MAX) { |
| /** |
| * Some of the platform may not support mcountinhibit. |
| * Checking the active_events is enough for them |
| */ |
| if (phs->active_events[cbase] != SBI_PMU_EVENT_IDX_INVALID) |
| continue; |
| /* If mcountinhibit is supported, the bit must be enabled */ |
| if ((sbi_hart_priv_version(scratch) >= SBI_HART_PRIV_VER_1_11) && |
| !__test_bit(cbase, &mctr_inhbt)) |
| continue; |
| /* We found a valid counter that is not started yet */ |
| ctr_idx = cbase; |
| } |
| } |
| |
| if (ctr_idx == SBI_ENOTSUPP) { |
| /** |
| * We can't find any programmable counters for cycle/instret. |
| * Return the fixed counter as they are mandatory anyways. |
| */ |
| if (fixed_ctr >= 0) |
| return pmu_fixed_ctr_update_inhibit_bits(fixed_ctr, flags); |
| else |
| return SBI_EFAIL; |
| } |
| ret = pmu_update_hw_mhpmevent(temp, ctr_idx, flags, event_idx, data); |
| |
| if (!ret) |
| ret = ctr_idx; |
| |
| return ret; |
| } |
| |
| |
| /** |
| * Any firmware counter can map to any firmware event. |
| * Thus, select the first available fw counter after sanity |
| * check. |
| */ |
| static int pmu_ctr_find_fw(struct sbi_pmu_hart_state *phs, |
| unsigned long cbase, unsigned long cmask, |
| uint32_t event_code, uint64_t edata) |
| { |
| int i, cidx; |
| |
| if ((event_code >= SBI_PMU_FW_MAX && |
| event_code <= SBI_PMU_FW_RESERVED_MAX) || |
| event_code > SBI_PMU_FW_PLATFORM) |
| return SBI_EINVAL; |
| |
| for_each_set_bit(i, &cmask, BITS_PER_LONG) { |
| cidx = i + cbase; |
| if (cidx < num_hw_ctrs || total_ctrs <= cidx) |
| continue; |
| if (phs->active_events[i] != SBI_PMU_EVENT_IDX_INVALID) |
| continue; |
| if (SBI_PMU_FW_PLATFORM == event_code && |
| pmu_dev && pmu_dev->fw_counter_match_encoding) { |
| if (!pmu_dev->fw_counter_match_encoding(phs->hartid, |
| cidx - num_hw_ctrs, |
| edata)) |
| continue; |
| } |
| |
| return i; |
| } |
| |
| return SBI_ENOTSUPP; |
| } |
| |
| int sbi_pmu_ctr_cfg_match(unsigned long cidx_base, unsigned long cidx_mask, |
| unsigned long flags, unsigned long event_idx, |
| uint64_t event_data) |
| { |
| struct sbi_pmu_hart_state *phs = pmu_thishart_state_ptr(); |
| int ret, event_type, ctr_idx = SBI_ENOTSUPP; |
| u32 event_code; |
| |
| /* Do a basic sanity check of counter base & mask */ |
| if ((cidx_base + sbi_fls(cidx_mask)) >= total_ctrs) |
| return SBI_EINVAL; |
| |
| event_type = pmu_event_validate(phs, event_idx, event_data); |
| if (event_type < 0) |
| return SBI_EINVAL; |
| event_code = get_cidx_code(event_idx); |
| |
| if (flags & SBI_PMU_CFG_FLAG_SKIP_MATCH) { |
| /* The caller wants to skip the match because it already knows the |
| * counter idx for the given event. Verify that the counter idx |
| * is still valid. |
| * As per the specification, we should "unconditionally select |
| * the first counter from the set of counters specified by the |
| * counter_idx_base and counter_idx_mask". |
| */ |
| unsigned long cidx_first = cidx_base + sbi_ffs(cidx_mask); |
| |
| if (phs->active_events[cidx_first] == SBI_PMU_EVENT_IDX_INVALID) |
| return SBI_EINVAL; |
| ctr_idx = cidx_first; |
| goto skip_match; |
| } |
| |
| if (event_type == SBI_PMU_EVENT_TYPE_FW) { |
| /* Any firmware counter can be used track any firmware event */ |
| ctr_idx = pmu_ctr_find_fw(phs, cidx_base, cidx_mask, |
| event_code, event_data); |
| if (event_code == SBI_PMU_FW_PLATFORM) |
| phs->fw_counters_data[ctr_idx - num_hw_ctrs] = |
| event_data; |
| } else { |
| ctr_idx = pmu_ctr_find_hw(phs, cidx_base, cidx_mask, flags, |
| event_idx, event_data); |
| } |
| |
| if (ctr_idx < 0) |
| return SBI_ENOTSUPP; |
| |
| phs->active_events[ctr_idx] = event_idx; |
| skip_match: |
| if (event_type == SBI_PMU_EVENT_TYPE_HW) { |
| if (flags & SBI_PMU_CFG_FLAG_CLEAR_VALUE) |
| pmu_ctr_write_hw(ctr_idx, 0); |
| if (flags & SBI_PMU_CFG_FLAG_AUTO_START) |
| pmu_ctr_start_hw(ctr_idx, 0, false); |
| } else if (event_type == SBI_PMU_EVENT_TYPE_FW) { |
| if (flags & SBI_PMU_CFG_FLAG_CLEAR_VALUE) |
| phs->fw_counters_data[ctr_idx - num_hw_ctrs] = 0; |
| if (flags & SBI_PMU_CFG_FLAG_AUTO_START) { |
| if (SBI_PMU_FW_PLATFORM == event_code && |
| pmu_dev && pmu_dev->fw_counter_start) { |
| ret = pmu_dev->fw_counter_start( |
| phs->hartid, |
| ctr_idx - num_hw_ctrs, event_data); |
| if (ret) |
| return ret; |
| } |
| phs->fw_counters_started |= BIT(ctr_idx - num_hw_ctrs); |
| } |
| } |
| |
| return ctr_idx; |
| } |
| |
| int sbi_pmu_ctr_incr_fw(enum sbi_pmu_fw_event_code_id fw_id) |
| { |
| u32 cidx; |
| uint64_t *fcounter = NULL; |
| struct sbi_pmu_hart_state *phs = pmu_thishart_state_ptr(); |
| |
| if (likely(!phs->fw_counters_started)) |
| return 0; |
| |
| if (unlikely(fw_id >= SBI_PMU_FW_MAX)) |
| return SBI_EINVAL; |
| |
| for (cidx = num_hw_ctrs; cidx < total_ctrs; cidx++) { |
| if (get_cidx_code(phs->active_events[cidx]) == fw_id && |
| (phs->fw_counters_started & BIT(cidx - num_hw_ctrs))) { |
| fcounter = &phs->fw_counters_data[cidx - num_hw_ctrs]; |
| break; |
| } |
| } |
| |
| if (fcounter) |
| (*fcounter)++; |
| |
| return 0; |
| } |
| |
| unsigned long sbi_pmu_num_ctr(void) |
| { |
| return (num_hw_ctrs + SBI_PMU_FW_CTR_MAX); |
| } |
| |
| int sbi_pmu_ctr_get_info(uint32_t cidx, unsigned long *ctr_info) |
| { |
| int width; |
| union sbi_pmu_ctr_info cinfo = {0}; |
| struct sbi_scratch *scratch = sbi_scratch_thishart_ptr(); |
| |
| /* Sanity check. Counter1 is not mapped at all */ |
| if (cidx >= total_ctrs || cidx == 1) |
| return SBI_EINVAL; |
| |
| /* We have 31 HW counters with 31 being the last index(MHPMCOUNTER31) */ |
| if (cidx < num_hw_ctrs) { |
| cinfo.type = SBI_PMU_CTR_TYPE_HW; |
| cinfo.csr = CSR_CYCLE + cidx; |
| /* mcycle & minstret are always 64 bit */ |
| if (cidx == 0 || cidx == 2) |
| cinfo.width = 63; |
| else |
| cinfo.width = sbi_hart_mhpm_bits(scratch) - 1; |
| } else { |
| /* it's a firmware counter */ |
| cinfo.type = SBI_PMU_CTR_TYPE_FW; |
| /* Firmware counters are always 64 bits wide */ |
| cinfo.width = 63; |
| if (pmu_dev && pmu_dev->fw_counter_width) { |
| width = pmu_dev->fw_counter_width(); |
| if (width) |
| cinfo.width = width - 1; |
| } |
| } |
| |
| *ctr_info = cinfo.value; |
| |
| return 0; |
| } |
| |
| static void pmu_reset_event_map(struct sbi_pmu_hart_state *phs) |
| { |
| int j; |
| |
| /* Initialize the counter to event mapping table */ |
| for (j = 3; j < total_ctrs; j++) |
| phs->active_events[j] = SBI_PMU_EVENT_IDX_INVALID; |
| for (j = 0; j < SBI_PMU_FW_CTR_MAX; j++) |
| phs->fw_counters_data[j] = 0; |
| phs->fw_counters_started = 0; |
| } |
| |
| const struct sbi_pmu_device *sbi_pmu_get_device(void) |
| { |
| return pmu_dev; |
| } |
| |
| void sbi_pmu_set_device(const struct sbi_pmu_device *dev) |
| { |
| if (!dev || pmu_dev) |
| return; |
| |
| pmu_dev = dev; |
| } |
| |
| void sbi_pmu_exit(struct sbi_scratch *scratch) |
| { |
| if (sbi_hart_priv_version(scratch) >= SBI_HART_PRIV_VER_1_11) |
| csr_write(CSR_MCOUNTINHIBIT, 0xFFFFFFF8); |
| |
| if (sbi_hart_priv_version(scratch) >= SBI_HART_PRIV_VER_1_10) |
| csr_write(CSR_MCOUNTEREN, -1); |
| |
| pmu_reset_event_map(pmu_get_hart_state_ptr(scratch)); |
| } |
| |
| int sbi_pmu_init(struct sbi_scratch *scratch, bool cold_boot) |
| { |
| int hpm_count = sbi_fls(sbi_hart_mhpm_mask(scratch)); |
| struct sbi_pmu_hart_state *phs; |
| const struct sbi_platform *plat; |
| |
| if (cold_boot) { |
| hw_event_map = sbi_calloc(sizeof(*hw_event_map), |
| SBI_PMU_HW_EVENT_MAX); |
| if (!hw_event_map) |
| return SBI_ENOMEM; |
| |
| phs_ptr_offset = sbi_scratch_alloc_type_offset(void *); |
| if (!phs_ptr_offset) { |
| sbi_free(hw_event_map); |
| return SBI_ENOMEM; |
| } |
| |
| plat = sbi_platform_ptr(scratch); |
| /* Initialize hw pmu events */ |
| sbi_platform_pmu_init(plat); |
| |
| /* mcycle & minstret is available always */ |
| if (!hpm_count) |
| /* Only CY, TM & IR are implemented in the hw */ |
| num_hw_ctrs = 3; |
| else |
| num_hw_ctrs = hpm_count + 1; |
| |
| if (num_hw_ctrs > SBI_PMU_HW_CTR_MAX) |
| return SBI_EINVAL; |
| |
| total_ctrs = num_hw_ctrs + SBI_PMU_FW_CTR_MAX; |
| } |
| |
| phs = pmu_get_hart_state_ptr(scratch); |
| if (!phs) { |
| phs = sbi_zalloc(sizeof(*phs)); |
| if (!phs) |
| return SBI_ENOMEM; |
| phs->hartid = current_hartid(); |
| pmu_set_hart_state_ptr(scratch, phs); |
| } |
| |
| pmu_reset_event_map(phs); |
| |
| /* First three counters are fixed by the priv spec and we enable it by default */ |
| phs->active_events[0] = (SBI_PMU_EVENT_TYPE_HW << SBI_PMU_EVENT_IDX_TYPE_OFFSET) | |
| SBI_PMU_HW_CPU_CYCLES; |
| phs->active_events[1] = SBI_PMU_EVENT_IDX_INVALID; |
| phs->active_events[2] = (SBI_PMU_EVENT_TYPE_HW << SBI_PMU_EVENT_IDX_TYPE_OFFSET) | |
| SBI_PMU_HW_INSTRUCTIONS; |
| |
| return 0; |
| } |