|  | // SPDX-License-Identifier: GPL-2.0-only | 
|  | /* | 
|  | * Resource Director Technology(RDT) | 
|  | * - Monitoring code | 
|  | * | 
|  | * Copyright (C) 2017 Intel Corporation | 
|  | * | 
|  | * Author: | 
|  | *    Vikas Shivappa <vikas.shivappa@intel.com> | 
|  | * | 
|  | * This replaces the cqm.c based on perf but we reuse a lot of | 
|  | * code and datastructures originally from Peter Zijlstra and Matt Fleming. | 
|  | * | 
|  | * More information about RDT be found in the Intel (R) x86 Architecture | 
|  | * Software Developer Manual June 2016, volume 3, section 17.17. | 
|  | */ | 
|  |  | 
|  | #define pr_fmt(fmt)	"resctrl: " fmt | 
|  |  | 
|  | #include <linux/cpu.h> | 
|  | #include <linux/resctrl.h> | 
|  | #include <linux/sizes.h> | 
|  | #include <linux/slab.h> | 
|  |  | 
|  | #include "internal.h" | 
|  |  | 
|  | #define CREATE_TRACE_POINTS | 
|  |  | 
|  | #include "monitor_trace.h" | 
|  |  | 
|  | /** | 
|  | * struct rmid_entry - dirty tracking for all RMID. | 
|  | * @closid:	The CLOSID for this entry. | 
|  | * @rmid:	The RMID for this entry. | 
|  | * @busy:	The number of domains with cached data using this RMID. | 
|  | * @list:	Member of the rmid_free_lru list when busy == 0. | 
|  | * | 
|  | * Depending on the architecture the correct monitor is accessed using | 
|  | * both @closid and @rmid, or @rmid only. | 
|  | * | 
|  | * Take the rdtgroup_mutex when accessing. | 
|  | */ | 
|  | struct rmid_entry { | 
|  | u32				closid; | 
|  | u32				rmid; | 
|  | int				busy; | 
|  | struct list_head		list; | 
|  | }; | 
|  |  | 
|  | /* | 
|  | * @rmid_free_lru - A least recently used list of free RMIDs | 
|  | *     These RMIDs are guaranteed to have an occupancy less than the | 
|  | *     threshold occupancy | 
|  | */ | 
|  | static LIST_HEAD(rmid_free_lru); | 
|  |  | 
|  | /* | 
|  | * @closid_num_dirty_rmid    The number of dirty RMID each CLOSID has. | 
|  | *     Only allocated when CONFIG_RESCTRL_RMID_DEPENDS_ON_CLOSID is defined. | 
|  | *     Indexed by CLOSID. Protected by rdtgroup_mutex. | 
|  | */ | 
|  | static u32 *closid_num_dirty_rmid; | 
|  |  | 
|  | /* | 
|  | * @rmid_limbo_count - count of currently unused but (potentially) | 
|  | *     dirty RMIDs. | 
|  | *     This counts RMIDs that no one is currently using but that | 
|  | *     may have a occupancy value > resctrl_rmid_realloc_threshold. User can | 
|  | *     change the threshold occupancy value. | 
|  | */ | 
|  | static unsigned int rmid_limbo_count; | 
|  |  | 
|  | /* | 
|  | * @rmid_entry - The entry in the limbo and free lists. | 
|  | */ | 
|  | static struct rmid_entry	*rmid_ptrs; | 
|  |  | 
|  | /* | 
|  | * This is the threshold cache occupancy in bytes at which we will consider an | 
|  | * RMID available for re-allocation. | 
|  | */ | 
|  | unsigned int resctrl_rmid_realloc_threshold; | 
|  |  | 
|  | /* | 
|  | * This is the maximum value for the reallocation threshold, in bytes. | 
|  | */ | 
|  | unsigned int resctrl_rmid_realloc_limit; | 
|  |  | 
|  | /* | 
|  | * x86 and arm64 differ in their handling of monitoring. | 
|  | * x86's RMID are independent numbers, there is only one source of traffic | 
|  | * with an RMID value of '1'. | 
|  | * arm64's PMG extends the PARTID/CLOSID space, there are multiple sources of | 
|  | * traffic with a PMG value of '1', one for each CLOSID, meaning the RMID | 
|  | * value is no longer unique. | 
|  | * To account for this, resctrl uses an index. On x86 this is just the RMID, | 
|  | * on arm64 it encodes the CLOSID and RMID. This gives a unique number. | 
|  | * | 
|  | * The domain's rmid_busy_llc and rmid_ptrs[] are sized by index. The arch code | 
|  | * must accept an attempt to read every index. | 
|  | */ | 
|  | static inline struct rmid_entry *__rmid_entry(u32 idx) | 
|  | { | 
|  | struct rmid_entry *entry; | 
|  | u32 closid, rmid; | 
|  |  | 
|  | entry = &rmid_ptrs[idx]; | 
|  | resctrl_arch_rmid_idx_decode(idx, &closid, &rmid); | 
|  |  | 
|  | WARN_ON_ONCE(entry->closid != closid); | 
|  | WARN_ON_ONCE(entry->rmid != rmid); | 
|  |  | 
|  | return entry; | 
|  | } | 
|  |  | 
|  | static void limbo_release_entry(struct rmid_entry *entry) | 
|  | { | 
|  | lockdep_assert_held(&rdtgroup_mutex); | 
|  |  | 
|  | rmid_limbo_count--; | 
|  | list_add_tail(&entry->list, &rmid_free_lru); | 
|  |  | 
|  | if (IS_ENABLED(CONFIG_RESCTRL_RMID_DEPENDS_ON_CLOSID)) | 
|  | closid_num_dirty_rmid[entry->closid]--; | 
|  | } | 
|  |  | 
|  | /* | 
|  | * Check the RMIDs that are marked as busy for this domain. If the | 
|  | * reported LLC occupancy is below the threshold clear the busy bit and | 
|  | * decrement the count. If the busy count gets to zero on an RMID, we | 
|  | * free the RMID | 
|  | */ | 
|  | void __check_limbo(struct rdt_mon_domain *d, bool force_free) | 
|  | { | 
|  | struct rdt_resource *r = resctrl_arch_get_resource(RDT_RESOURCE_L3); | 
|  | u32 idx_limit = resctrl_arch_system_num_rmid_idx(); | 
|  | struct rmid_entry *entry; | 
|  | u32 idx, cur_idx = 1; | 
|  | void *arch_mon_ctx; | 
|  | bool rmid_dirty; | 
|  | u64 val = 0; | 
|  |  | 
|  | arch_mon_ctx = resctrl_arch_mon_ctx_alloc(r, QOS_L3_OCCUP_EVENT_ID); | 
|  | if (IS_ERR(arch_mon_ctx)) { | 
|  | pr_warn_ratelimited("Failed to allocate monitor context: %ld", | 
|  | PTR_ERR(arch_mon_ctx)); | 
|  | return; | 
|  | } | 
|  |  | 
|  | /* | 
|  | * Skip RMID 0 and start from RMID 1 and check all the RMIDs that | 
|  | * are marked as busy for occupancy < threshold. If the occupancy | 
|  | * is less than the threshold decrement the busy counter of the | 
|  | * RMID and move it to the free list when the counter reaches 0. | 
|  | */ | 
|  | for (;;) { | 
|  | idx = find_next_bit(d->rmid_busy_llc, idx_limit, cur_idx); | 
|  | if (idx >= idx_limit) | 
|  | break; | 
|  |  | 
|  | entry = __rmid_entry(idx); | 
|  | if (resctrl_arch_rmid_read(r, d, entry->closid, entry->rmid, | 
|  | QOS_L3_OCCUP_EVENT_ID, &val, | 
|  | arch_mon_ctx)) { | 
|  | rmid_dirty = true; | 
|  | } else { | 
|  | rmid_dirty = (val >= resctrl_rmid_realloc_threshold); | 
|  |  | 
|  | /* | 
|  | * x86's CLOSID and RMID are independent numbers, so the entry's | 
|  | * CLOSID is an empty CLOSID (X86_RESCTRL_EMPTY_CLOSID). On Arm the | 
|  | * RMID (PMG) extends the CLOSID (PARTID) space with bits that aren't | 
|  | * used to select the configuration. It is thus necessary to track both | 
|  | * CLOSID and RMID because there may be dependencies between them | 
|  | * on some architectures. | 
|  | */ | 
|  | trace_mon_llc_occupancy_limbo(entry->closid, entry->rmid, d->hdr.id, val); | 
|  | } | 
|  |  | 
|  | if (force_free || !rmid_dirty) { | 
|  | clear_bit(idx, d->rmid_busy_llc); | 
|  | if (!--entry->busy) | 
|  | limbo_release_entry(entry); | 
|  | } | 
|  | cur_idx = idx + 1; | 
|  | } | 
|  |  | 
|  | resctrl_arch_mon_ctx_free(r, QOS_L3_OCCUP_EVENT_ID, arch_mon_ctx); | 
|  | } | 
|  |  | 
|  | bool has_busy_rmid(struct rdt_mon_domain *d) | 
|  | { | 
|  | u32 idx_limit = resctrl_arch_system_num_rmid_idx(); | 
|  |  | 
|  | return find_first_bit(d->rmid_busy_llc, idx_limit) != idx_limit; | 
|  | } | 
|  |  | 
|  | static struct rmid_entry *resctrl_find_free_rmid(u32 closid) | 
|  | { | 
|  | struct rmid_entry *itr; | 
|  | u32 itr_idx, cmp_idx; | 
|  |  | 
|  | if (list_empty(&rmid_free_lru)) | 
|  | return rmid_limbo_count ? ERR_PTR(-EBUSY) : ERR_PTR(-ENOSPC); | 
|  |  | 
|  | list_for_each_entry(itr, &rmid_free_lru, list) { | 
|  | /* | 
|  | * Get the index of this free RMID, and the index it would need | 
|  | * to be if it were used with this CLOSID. | 
|  | * If the CLOSID is irrelevant on this architecture, the two | 
|  | * index values are always the same on every entry and thus the | 
|  | * very first entry will be returned. | 
|  | */ | 
|  | itr_idx = resctrl_arch_rmid_idx_encode(itr->closid, itr->rmid); | 
|  | cmp_idx = resctrl_arch_rmid_idx_encode(closid, itr->rmid); | 
|  |  | 
|  | if (itr_idx == cmp_idx) | 
|  | return itr; | 
|  | } | 
|  |  | 
|  | return ERR_PTR(-ENOSPC); | 
|  | } | 
|  |  | 
|  | /** | 
|  | * resctrl_find_cleanest_closid() - Find a CLOSID where all the associated | 
|  | *                                  RMID are clean, or the CLOSID that has | 
|  | *                                  the most clean RMID. | 
|  | * | 
|  | * MPAM's equivalent of RMID are per-CLOSID, meaning a freshly allocated CLOSID | 
|  | * may not be able to allocate clean RMID. To avoid this the allocator will | 
|  | * choose the CLOSID with the most clean RMID. | 
|  | * | 
|  | * When the CLOSID and RMID are independent numbers, the first free CLOSID will | 
|  | * be returned. | 
|  | */ | 
|  | int resctrl_find_cleanest_closid(void) | 
|  | { | 
|  | u32 cleanest_closid = ~0; | 
|  | int i = 0; | 
|  |  | 
|  | lockdep_assert_held(&rdtgroup_mutex); | 
|  |  | 
|  | if (!IS_ENABLED(CONFIG_RESCTRL_RMID_DEPENDS_ON_CLOSID)) | 
|  | return -EIO; | 
|  |  | 
|  | for (i = 0; i < closids_supported(); i++) { | 
|  | int num_dirty; | 
|  |  | 
|  | if (closid_allocated(i)) | 
|  | continue; | 
|  |  | 
|  | num_dirty = closid_num_dirty_rmid[i]; | 
|  | if (num_dirty == 0) | 
|  | return i; | 
|  |  | 
|  | if (cleanest_closid == ~0) | 
|  | cleanest_closid = i; | 
|  |  | 
|  | if (num_dirty < closid_num_dirty_rmid[cleanest_closid]) | 
|  | cleanest_closid = i; | 
|  | } | 
|  |  | 
|  | if (cleanest_closid == ~0) | 
|  | return -ENOSPC; | 
|  |  | 
|  | return cleanest_closid; | 
|  | } | 
|  |  | 
|  | /* | 
|  | * For MPAM the RMID value is not unique, and has to be considered with | 
|  | * the CLOSID. The (CLOSID, RMID) pair is allocated on all domains, which | 
|  | * allows all domains to be managed by a single free list. | 
|  | * Each domain also has a rmid_busy_llc to reduce the work of the limbo handler. | 
|  | */ | 
|  | int alloc_rmid(u32 closid) | 
|  | { | 
|  | struct rmid_entry *entry; | 
|  |  | 
|  | lockdep_assert_held(&rdtgroup_mutex); | 
|  |  | 
|  | entry = resctrl_find_free_rmid(closid); | 
|  | if (IS_ERR(entry)) | 
|  | return PTR_ERR(entry); | 
|  |  | 
|  | list_del(&entry->list); | 
|  | return entry->rmid; | 
|  | } | 
|  |  | 
|  | static void add_rmid_to_limbo(struct rmid_entry *entry) | 
|  | { | 
|  | struct rdt_resource *r = resctrl_arch_get_resource(RDT_RESOURCE_L3); | 
|  | struct rdt_mon_domain *d; | 
|  | u32 idx; | 
|  |  | 
|  | lockdep_assert_held(&rdtgroup_mutex); | 
|  |  | 
|  | /* Walking r->domains, ensure it can't race with cpuhp */ | 
|  | lockdep_assert_cpus_held(); | 
|  |  | 
|  | idx = resctrl_arch_rmid_idx_encode(entry->closid, entry->rmid); | 
|  |  | 
|  | entry->busy = 0; | 
|  | list_for_each_entry(d, &r->mon_domains, hdr.list) { | 
|  | /* | 
|  | * For the first limbo RMID in the domain, | 
|  | * setup up the limbo worker. | 
|  | */ | 
|  | if (!has_busy_rmid(d)) | 
|  | cqm_setup_limbo_handler(d, CQM_LIMBOCHECK_INTERVAL, | 
|  | RESCTRL_PICK_ANY_CPU); | 
|  | set_bit(idx, d->rmid_busy_llc); | 
|  | entry->busy++; | 
|  | } | 
|  |  | 
|  | rmid_limbo_count++; | 
|  | if (IS_ENABLED(CONFIG_RESCTRL_RMID_DEPENDS_ON_CLOSID)) | 
|  | closid_num_dirty_rmid[entry->closid]++; | 
|  | } | 
|  |  | 
|  | void free_rmid(u32 closid, u32 rmid) | 
|  | { | 
|  | u32 idx = resctrl_arch_rmid_idx_encode(closid, rmid); | 
|  | struct rmid_entry *entry; | 
|  |  | 
|  | lockdep_assert_held(&rdtgroup_mutex); | 
|  |  | 
|  | /* | 
|  | * Do not allow the default rmid to be free'd. Comparing by index | 
|  | * allows architectures that ignore the closid parameter to avoid an | 
|  | * unnecessary check. | 
|  | */ | 
|  | if (!resctrl_arch_mon_capable() || | 
|  | idx == resctrl_arch_rmid_idx_encode(RESCTRL_RESERVED_CLOSID, | 
|  | RESCTRL_RESERVED_RMID)) | 
|  | return; | 
|  |  | 
|  | entry = __rmid_entry(idx); | 
|  |  | 
|  | if (resctrl_is_mon_event_enabled(QOS_L3_OCCUP_EVENT_ID)) | 
|  | add_rmid_to_limbo(entry); | 
|  | else | 
|  | list_add_tail(&entry->list, &rmid_free_lru); | 
|  | } | 
|  |  | 
|  | static struct mbm_state *get_mbm_state(struct rdt_mon_domain *d, u32 closid, | 
|  | u32 rmid, enum resctrl_event_id evtid) | 
|  | { | 
|  | u32 idx = resctrl_arch_rmid_idx_encode(closid, rmid); | 
|  | struct mbm_state *state; | 
|  |  | 
|  | if (!resctrl_is_mbm_event(evtid)) | 
|  | return NULL; | 
|  |  | 
|  | state = d->mbm_states[MBM_STATE_IDX(evtid)]; | 
|  |  | 
|  | return state ? &state[idx] : NULL; | 
|  | } | 
|  |  | 
|  | /* | 
|  | * mbm_cntr_get() - Return the counter ID for the matching @evtid and @rdtgrp. | 
|  | * | 
|  | * Return: | 
|  | * Valid counter ID on success, or -ENOENT on failure. | 
|  | */ | 
|  | static int mbm_cntr_get(struct rdt_resource *r, struct rdt_mon_domain *d, | 
|  | struct rdtgroup *rdtgrp, enum resctrl_event_id evtid) | 
|  | { | 
|  | int cntr_id; | 
|  |  | 
|  | if (!r->mon.mbm_cntr_assignable) | 
|  | return -ENOENT; | 
|  |  | 
|  | if (!resctrl_is_mbm_event(evtid)) | 
|  | return -ENOENT; | 
|  |  | 
|  | for (cntr_id = 0; cntr_id < r->mon.num_mbm_cntrs; cntr_id++) { | 
|  | if (d->cntr_cfg[cntr_id].rdtgrp == rdtgrp && | 
|  | d->cntr_cfg[cntr_id].evtid == evtid) | 
|  | return cntr_id; | 
|  | } | 
|  |  | 
|  | return -ENOENT; | 
|  | } | 
|  |  | 
|  | /* | 
|  | * mbm_cntr_alloc() - Initialize and return a new counter ID in the domain @d. | 
|  | * Caller must ensure that the specified event is not assigned already. | 
|  | * | 
|  | * Return: | 
|  | * Valid counter ID on success, or -ENOSPC on failure. | 
|  | */ | 
|  | static int mbm_cntr_alloc(struct rdt_resource *r, struct rdt_mon_domain *d, | 
|  | struct rdtgroup *rdtgrp, enum resctrl_event_id evtid) | 
|  | { | 
|  | int cntr_id; | 
|  |  | 
|  | for (cntr_id = 0; cntr_id < r->mon.num_mbm_cntrs; cntr_id++) { | 
|  | if (!d->cntr_cfg[cntr_id].rdtgrp) { | 
|  | d->cntr_cfg[cntr_id].rdtgrp = rdtgrp; | 
|  | d->cntr_cfg[cntr_id].evtid = evtid; | 
|  | return cntr_id; | 
|  | } | 
|  | } | 
|  |  | 
|  | return -ENOSPC; | 
|  | } | 
|  |  | 
|  | /* | 
|  | * mbm_cntr_free() - Clear the counter ID configuration details in the domain @d. | 
|  | */ | 
|  | static void mbm_cntr_free(struct rdt_mon_domain *d, int cntr_id) | 
|  | { | 
|  | memset(&d->cntr_cfg[cntr_id], 0, sizeof(*d->cntr_cfg)); | 
|  | } | 
|  |  | 
|  | static int __mon_event_count(struct rdtgroup *rdtgrp, struct rmid_read *rr) | 
|  | { | 
|  | int cpu = smp_processor_id(); | 
|  | u32 closid = rdtgrp->closid; | 
|  | u32 rmid = rdtgrp->mon.rmid; | 
|  | struct rdt_mon_domain *d; | 
|  | int cntr_id = -ENOENT; | 
|  | struct mbm_state *m; | 
|  | int err, ret; | 
|  | u64 tval = 0; | 
|  |  | 
|  | if (rr->is_mbm_cntr) { | 
|  | cntr_id = mbm_cntr_get(rr->r, rr->d, rdtgrp, rr->evtid); | 
|  | if (cntr_id < 0) { | 
|  | rr->err = -ENOENT; | 
|  | return -EINVAL; | 
|  | } | 
|  | } | 
|  |  | 
|  | if (rr->first) { | 
|  | if (rr->is_mbm_cntr) | 
|  | resctrl_arch_reset_cntr(rr->r, rr->d, closid, rmid, cntr_id, rr->evtid); | 
|  | else | 
|  | resctrl_arch_reset_rmid(rr->r, rr->d, closid, rmid, rr->evtid); | 
|  | m = get_mbm_state(rr->d, closid, rmid, rr->evtid); | 
|  | if (m) | 
|  | memset(m, 0, sizeof(struct mbm_state)); | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | if (rr->d) { | 
|  | /* Reading a single domain, must be on a CPU in that domain. */ | 
|  | if (!cpumask_test_cpu(cpu, &rr->d->hdr.cpu_mask)) | 
|  | return -EINVAL; | 
|  | if (rr->is_mbm_cntr) | 
|  | rr->err = resctrl_arch_cntr_read(rr->r, rr->d, closid, rmid, cntr_id, | 
|  | rr->evtid, &tval); | 
|  | else | 
|  | rr->err = resctrl_arch_rmid_read(rr->r, rr->d, closid, rmid, | 
|  | rr->evtid, &tval, rr->arch_mon_ctx); | 
|  | if (rr->err) | 
|  | return rr->err; | 
|  |  | 
|  | rr->val += tval; | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | /* Summing domains that share a cache, must be on a CPU for that cache. */ | 
|  | if (!cpumask_test_cpu(cpu, &rr->ci->shared_cpu_map)) | 
|  | return -EINVAL; | 
|  |  | 
|  | /* | 
|  | * Legacy files must report the sum of an event across all | 
|  | * domains that share the same L3 cache instance. | 
|  | * Report success if a read from any domain succeeds, -EINVAL | 
|  | * (translated to "Unavailable" for user space) if reading from | 
|  | * all domains fail for any reason. | 
|  | */ | 
|  | ret = -EINVAL; | 
|  | list_for_each_entry(d, &rr->r->mon_domains, hdr.list) { | 
|  | if (d->ci_id != rr->ci->id) | 
|  | continue; | 
|  | if (rr->is_mbm_cntr) | 
|  | err = resctrl_arch_cntr_read(rr->r, d, closid, rmid, cntr_id, | 
|  | rr->evtid, &tval); | 
|  | else | 
|  | err = resctrl_arch_rmid_read(rr->r, d, closid, rmid, | 
|  | rr->evtid, &tval, rr->arch_mon_ctx); | 
|  | if (!err) { | 
|  | rr->val += tval; | 
|  | ret = 0; | 
|  | } | 
|  | } | 
|  |  | 
|  | if (ret) | 
|  | rr->err = ret; | 
|  |  | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | /* | 
|  | * mbm_bw_count() - Update bw count from values previously read by | 
|  | *		    __mon_event_count(). | 
|  | * @rdtgrp:	resctrl group associated with the CLOSID and RMID to identify | 
|  | *		the cached mbm_state. | 
|  | * @rr:		The struct rmid_read populated by __mon_event_count(). | 
|  | * | 
|  | * Supporting function to calculate the memory bandwidth | 
|  | * and delta bandwidth in MBps. The chunks value previously read by | 
|  | * __mon_event_count() is compared with the chunks value from the previous | 
|  | * invocation. This must be called once per second to maintain values in MBps. | 
|  | */ | 
|  | static void mbm_bw_count(struct rdtgroup *rdtgrp, struct rmid_read *rr) | 
|  | { | 
|  | u64 cur_bw, bytes, cur_bytes; | 
|  | u32 closid = rdtgrp->closid; | 
|  | u32 rmid = rdtgrp->mon.rmid; | 
|  | struct mbm_state *m; | 
|  |  | 
|  | m = get_mbm_state(rr->d, closid, rmid, rr->evtid); | 
|  | if (WARN_ON_ONCE(!m)) | 
|  | return; | 
|  |  | 
|  | cur_bytes = rr->val; | 
|  | bytes = cur_bytes - m->prev_bw_bytes; | 
|  | m->prev_bw_bytes = cur_bytes; | 
|  |  | 
|  | cur_bw = bytes / SZ_1M; | 
|  |  | 
|  | m->prev_bw = cur_bw; | 
|  | } | 
|  |  | 
|  | /* | 
|  | * This is scheduled by mon_event_read() to read the CQM/MBM counters | 
|  | * on a domain. | 
|  | */ | 
|  | void mon_event_count(void *info) | 
|  | { | 
|  | struct rdtgroup *rdtgrp, *entry; | 
|  | struct rmid_read *rr = info; | 
|  | struct list_head *head; | 
|  | int ret; | 
|  |  | 
|  | rdtgrp = rr->rgrp; | 
|  |  | 
|  | ret = __mon_event_count(rdtgrp, rr); | 
|  |  | 
|  | /* | 
|  | * For Ctrl groups read data from child monitor groups and | 
|  | * add them together. Count events which are read successfully. | 
|  | * Discard the rmid_read's reporting errors. | 
|  | */ | 
|  | head = &rdtgrp->mon.crdtgrp_list; | 
|  |  | 
|  | if (rdtgrp->type == RDTCTRL_GROUP) { | 
|  | list_for_each_entry(entry, head, mon.crdtgrp_list) { | 
|  | if (__mon_event_count(entry, rr) == 0) | 
|  | ret = 0; | 
|  | } | 
|  | } | 
|  |  | 
|  | /* | 
|  | * __mon_event_count() calls for newly created monitor groups may | 
|  | * report -EINVAL/Unavailable if the monitor hasn't seen any traffic. | 
|  | * Discard error if any of the monitor event reads succeeded. | 
|  | */ | 
|  | if (ret == 0) | 
|  | rr->err = 0; | 
|  | } | 
|  |  | 
|  | static struct rdt_ctrl_domain *get_ctrl_domain_from_cpu(int cpu, | 
|  | struct rdt_resource *r) | 
|  | { | 
|  | struct rdt_ctrl_domain *d; | 
|  |  | 
|  | lockdep_assert_cpus_held(); | 
|  |  | 
|  | list_for_each_entry(d, &r->ctrl_domains, hdr.list) { | 
|  | /* Find the domain that contains this CPU */ | 
|  | if (cpumask_test_cpu(cpu, &d->hdr.cpu_mask)) | 
|  | return d; | 
|  | } | 
|  |  | 
|  | return NULL; | 
|  | } | 
|  |  | 
|  | /* | 
|  | * Feedback loop for MBA software controller (mba_sc) | 
|  | * | 
|  | * mba_sc is a feedback loop where we periodically read MBM counters and | 
|  | * adjust the bandwidth percentage values via the IA32_MBA_THRTL_MSRs so | 
|  | * that: | 
|  | * | 
|  | *   current bandwidth(cur_bw) < user specified bandwidth(user_bw) | 
|  | * | 
|  | * This uses the MBM counters to measure the bandwidth and MBA throttle | 
|  | * MSRs to control the bandwidth for a particular rdtgrp. It builds on the | 
|  | * fact that resctrl rdtgroups have both monitoring and control. | 
|  | * | 
|  | * The frequency of the checks is 1s and we just tag along the MBM overflow | 
|  | * timer. Having 1s interval makes the calculation of bandwidth simpler. | 
|  | * | 
|  | * Although MBA's goal is to restrict the bandwidth to a maximum, there may | 
|  | * be a need to increase the bandwidth to avoid unnecessarily restricting | 
|  | * the L2 <-> L3 traffic. | 
|  | * | 
|  | * Since MBA controls the L2 external bandwidth where as MBM measures the | 
|  | * L3 external bandwidth the following sequence could lead to such a | 
|  | * situation. | 
|  | * | 
|  | * Consider an rdtgroup which had high L3 <-> memory traffic in initial | 
|  | * phases -> mba_sc kicks in and reduced bandwidth percentage values -> but | 
|  | * after some time rdtgroup has mostly L2 <-> L3 traffic. | 
|  | * | 
|  | * In this case we may restrict the rdtgroup's L2 <-> L3 traffic as its | 
|  | * throttle MSRs already have low percentage values.  To avoid | 
|  | * unnecessarily restricting such rdtgroups, we also increase the bandwidth. | 
|  | */ | 
|  | static void update_mba_bw(struct rdtgroup *rgrp, struct rdt_mon_domain *dom_mbm) | 
|  | { | 
|  | u32 closid, rmid, cur_msr_val, new_msr_val; | 
|  | struct mbm_state *pmbm_data, *cmbm_data; | 
|  | struct rdt_ctrl_domain *dom_mba; | 
|  | enum resctrl_event_id evt_id; | 
|  | struct rdt_resource *r_mba; | 
|  | struct list_head *head; | 
|  | struct rdtgroup *entry; | 
|  | u32 cur_bw, user_bw; | 
|  |  | 
|  | r_mba = resctrl_arch_get_resource(RDT_RESOURCE_MBA); | 
|  | evt_id = rgrp->mba_mbps_event; | 
|  |  | 
|  | closid = rgrp->closid; | 
|  | rmid = rgrp->mon.rmid; | 
|  | pmbm_data = get_mbm_state(dom_mbm, closid, rmid, evt_id); | 
|  | if (WARN_ON_ONCE(!pmbm_data)) | 
|  | return; | 
|  |  | 
|  | dom_mba = get_ctrl_domain_from_cpu(smp_processor_id(), r_mba); | 
|  | if (!dom_mba) { | 
|  | pr_warn_once("Failure to get domain for MBA update\n"); | 
|  | return; | 
|  | } | 
|  |  | 
|  | cur_bw = pmbm_data->prev_bw; | 
|  | user_bw = dom_mba->mbps_val[closid]; | 
|  |  | 
|  | /* MBA resource doesn't support CDP */ | 
|  | cur_msr_val = resctrl_arch_get_config(r_mba, dom_mba, closid, CDP_NONE); | 
|  |  | 
|  | /* | 
|  | * For Ctrl groups read data from child monitor groups. | 
|  | */ | 
|  | head = &rgrp->mon.crdtgrp_list; | 
|  | list_for_each_entry(entry, head, mon.crdtgrp_list) { | 
|  | cmbm_data = get_mbm_state(dom_mbm, entry->closid, entry->mon.rmid, evt_id); | 
|  | if (WARN_ON_ONCE(!cmbm_data)) | 
|  | return; | 
|  | cur_bw += cmbm_data->prev_bw; | 
|  | } | 
|  |  | 
|  | /* | 
|  | * Scale up/down the bandwidth linearly for the ctrl group.  The | 
|  | * bandwidth step is the bandwidth granularity specified by the | 
|  | * hardware. | 
|  | * Always increase throttling if current bandwidth is above the | 
|  | * target set by user. | 
|  | * But avoid thrashing up and down on every poll by checking | 
|  | * whether a decrease in throttling is likely to push the group | 
|  | * back over target. E.g. if currently throttling to 30% of bandwidth | 
|  | * on a system with 10% granularity steps, check whether moving to | 
|  | * 40% would go past the limit by multiplying current bandwidth by | 
|  | * "(30 + 10) / 30". | 
|  | */ | 
|  | if (cur_msr_val > r_mba->membw.min_bw && user_bw < cur_bw) { | 
|  | new_msr_val = cur_msr_val - r_mba->membw.bw_gran; | 
|  | } else if (cur_msr_val < MAX_MBA_BW && | 
|  | (user_bw > (cur_bw * (cur_msr_val + r_mba->membw.min_bw) / cur_msr_val))) { | 
|  | new_msr_val = cur_msr_val + r_mba->membw.bw_gran; | 
|  | } else { | 
|  | return; | 
|  | } | 
|  |  | 
|  | resctrl_arch_update_one(r_mba, dom_mba, closid, CDP_NONE, new_msr_val); | 
|  | } | 
|  |  | 
|  | static void mbm_update_one_event(struct rdt_resource *r, struct rdt_mon_domain *d, | 
|  | struct rdtgroup *rdtgrp, enum resctrl_event_id evtid) | 
|  | { | 
|  | struct rmid_read rr = {0}; | 
|  |  | 
|  | rr.r = r; | 
|  | rr.d = d; | 
|  | rr.evtid = evtid; | 
|  | if (resctrl_arch_mbm_cntr_assign_enabled(r)) { | 
|  | rr.is_mbm_cntr = true; | 
|  | } else { | 
|  | rr.arch_mon_ctx = resctrl_arch_mon_ctx_alloc(rr.r, rr.evtid); | 
|  | if (IS_ERR(rr.arch_mon_ctx)) { | 
|  | pr_warn_ratelimited("Failed to allocate monitor context: %ld", | 
|  | PTR_ERR(rr.arch_mon_ctx)); | 
|  | return; | 
|  | } | 
|  | } | 
|  |  | 
|  | __mon_event_count(rdtgrp, &rr); | 
|  |  | 
|  | /* | 
|  | * If the software controller is enabled, compute the | 
|  | * bandwidth for this event id. | 
|  | */ | 
|  | if (is_mba_sc(NULL)) | 
|  | mbm_bw_count(rdtgrp, &rr); | 
|  |  | 
|  | if (rr.arch_mon_ctx) | 
|  | resctrl_arch_mon_ctx_free(rr.r, rr.evtid, rr.arch_mon_ctx); | 
|  | } | 
|  |  | 
|  | static void mbm_update(struct rdt_resource *r, struct rdt_mon_domain *d, | 
|  | struct rdtgroup *rdtgrp) | 
|  | { | 
|  | /* | 
|  | * This is protected from concurrent reads from user as both | 
|  | * the user and overflow handler hold the global mutex. | 
|  | */ | 
|  | if (resctrl_is_mon_event_enabled(QOS_L3_MBM_TOTAL_EVENT_ID)) | 
|  | mbm_update_one_event(r, d, rdtgrp, QOS_L3_MBM_TOTAL_EVENT_ID); | 
|  |  | 
|  | if (resctrl_is_mon_event_enabled(QOS_L3_MBM_LOCAL_EVENT_ID)) | 
|  | mbm_update_one_event(r, d, rdtgrp, QOS_L3_MBM_LOCAL_EVENT_ID); | 
|  | } | 
|  |  | 
|  | /* | 
|  | * Handler to scan the limbo list and move the RMIDs | 
|  | * to free list whose occupancy < threshold_occupancy. | 
|  | */ | 
|  | void cqm_handle_limbo(struct work_struct *work) | 
|  | { | 
|  | unsigned long delay = msecs_to_jiffies(CQM_LIMBOCHECK_INTERVAL); | 
|  | struct rdt_mon_domain *d; | 
|  |  | 
|  | cpus_read_lock(); | 
|  | mutex_lock(&rdtgroup_mutex); | 
|  |  | 
|  | d = container_of(work, struct rdt_mon_domain, cqm_limbo.work); | 
|  |  | 
|  | __check_limbo(d, false); | 
|  |  | 
|  | if (has_busy_rmid(d)) { | 
|  | d->cqm_work_cpu = cpumask_any_housekeeping(&d->hdr.cpu_mask, | 
|  | RESCTRL_PICK_ANY_CPU); | 
|  | schedule_delayed_work_on(d->cqm_work_cpu, &d->cqm_limbo, | 
|  | delay); | 
|  | } | 
|  |  | 
|  | mutex_unlock(&rdtgroup_mutex); | 
|  | cpus_read_unlock(); | 
|  | } | 
|  |  | 
|  | /** | 
|  | * cqm_setup_limbo_handler() - Schedule the limbo handler to run for this | 
|  | *                             domain. | 
|  | * @dom:           The domain the limbo handler should run for. | 
|  | * @delay_ms:      How far in the future the handler should run. | 
|  | * @exclude_cpu:   Which CPU the handler should not run on, | 
|  | *		   RESCTRL_PICK_ANY_CPU to pick any CPU. | 
|  | */ | 
|  | void cqm_setup_limbo_handler(struct rdt_mon_domain *dom, unsigned long delay_ms, | 
|  | int exclude_cpu) | 
|  | { | 
|  | unsigned long delay = msecs_to_jiffies(delay_ms); | 
|  | int cpu; | 
|  |  | 
|  | cpu = cpumask_any_housekeeping(&dom->hdr.cpu_mask, exclude_cpu); | 
|  | dom->cqm_work_cpu = cpu; | 
|  |  | 
|  | if (cpu < nr_cpu_ids) | 
|  | schedule_delayed_work_on(cpu, &dom->cqm_limbo, delay); | 
|  | } | 
|  |  | 
|  | void mbm_handle_overflow(struct work_struct *work) | 
|  | { | 
|  | unsigned long delay = msecs_to_jiffies(MBM_OVERFLOW_INTERVAL); | 
|  | struct rdtgroup *prgrp, *crgrp; | 
|  | struct rdt_mon_domain *d; | 
|  | struct list_head *head; | 
|  | struct rdt_resource *r; | 
|  |  | 
|  | cpus_read_lock(); | 
|  | mutex_lock(&rdtgroup_mutex); | 
|  |  | 
|  | /* | 
|  | * If the filesystem has been unmounted this work no longer needs to | 
|  | * run. | 
|  | */ | 
|  | if (!resctrl_mounted || !resctrl_arch_mon_capable()) | 
|  | goto out_unlock; | 
|  |  | 
|  | r = resctrl_arch_get_resource(RDT_RESOURCE_L3); | 
|  | d = container_of(work, struct rdt_mon_domain, mbm_over.work); | 
|  |  | 
|  | list_for_each_entry(prgrp, &rdt_all_groups, rdtgroup_list) { | 
|  | mbm_update(r, d, prgrp); | 
|  |  | 
|  | head = &prgrp->mon.crdtgrp_list; | 
|  | list_for_each_entry(crgrp, head, mon.crdtgrp_list) | 
|  | mbm_update(r, d, crgrp); | 
|  |  | 
|  | if (is_mba_sc(NULL)) | 
|  | update_mba_bw(prgrp, d); | 
|  | } | 
|  |  | 
|  | /* | 
|  | * Re-check for housekeeping CPUs. This allows the overflow handler to | 
|  | * move off a nohz_full CPU quickly. | 
|  | */ | 
|  | d->mbm_work_cpu = cpumask_any_housekeeping(&d->hdr.cpu_mask, | 
|  | RESCTRL_PICK_ANY_CPU); | 
|  | schedule_delayed_work_on(d->mbm_work_cpu, &d->mbm_over, delay); | 
|  |  | 
|  | out_unlock: | 
|  | mutex_unlock(&rdtgroup_mutex); | 
|  | cpus_read_unlock(); | 
|  | } | 
|  |  | 
|  | /** | 
|  | * mbm_setup_overflow_handler() - Schedule the overflow handler to run for this | 
|  | *                                domain. | 
|  | * @dom:           The domain the overflow handler should run for. | 
|  | * @delay_ms:      How far in the future the handler should run. | 
|  | * @exclude_cpu:   Which CPU the handler should not run on, | 
|  | *		   RESCTRL_PICK_ANY_CPU to pick any CPU. | 
|  | */ | 
|  | void mbm_setup_overflow_handler(struct rdt_mon_domain *dom, unsigned long delay_ms, | 
|  | int exclude_cpu) | 
|  | { | 
|  | unsigned long delay = msecs_to_jiffies(delay_ms); | 
|  | int cpu; | 
|  |  | 
|  | /* | 
|  | * When a domain comes online there is no guarantee the filesystem is | 
|  | * mounted. If not, there is no need to catch counter overflow. | 
|  | */ | 
|  | if (!resctrl_mounted || !resctrl_arch_mon_capable()) | 
|  | return; | 
|  | cpu = cpumask_any_housekeeping(&dom->hdr.cpu_mask, exclude_cpu); | 
|  | dom->mbm_work_cpu = cpu; | 
|  |  | 
|  | if (cpu < nr_cpu_ids) | 
|  | schedule_delayed_work_on(cpu, &dom->mbm_over, delay); | 
|  | } | 
|  |  | 
|  | static int dom_data_init(struct rdt_resource *r) | 
|  | { | 
|  | u32 idx_limit = resctrl_arch_system_num_rmid_idx(); | 
|  | u32 num_closid = resctrl_arch_get_num_closid(r); | 
|  | struct rmid_entry *entry = NULL; | 
|  | int err = 0, i; | 
|  | u32 idx; | 
|  |  | 
|  | mutex_lock(&rdtgroup_mutex); | 
|  | if (IS_ENABLED(CONFIG_RESCTRL_RMID_DEPENDS_ON_CLOSID)) { | 
|  | u32 *tmp; | 
|  |  | 
|  | /* | 
|  | * If the architecture hasn't provided a sanitised value here, | 
|  | * this may result in larger arrays than necessary. Resctrl will | 
|  | * use a smaller system wide value based on the resources in | 
|  | * use. | 
|  | */ | 
|  | tmp = kcalloc(num_closid, sizeof(*tmp), GFP_KERNEL); | 
|  | if (!tmp) { | 
|  | err = -ENOMEM; | 
|  | goto out_unlock; | 
|  | } | 
|  |  | 
|  | closid_num_dirty_rmid = tmp; | 
|  | } | 
|  |  | 
|  | rmid_ptrs = kcalloc(idx_limit, sizeof(struct rmid_entry), GFP_KERNEL); | 
|  | if (!rmid_ptrs) { | 
|  | if (IS_ENABLED(CONFIG_RESCTRL_RMID_DEPENDS_ON_CLOSID)) { | 
|  | kfree(closid_num_dirty_rmid); | 
|  | closid_num_dirty_rmid = NULL; | 
|  | } | 
|  | err = -ENOMEM; | 
|  | goto out_unlock; | 
|  | } | 
|  |  | 
|  | for (i = 0; i < idx_limit; i++) { | 
|  | entry = &rmid_ptrs[i]; | 
|  | INIT_LIST_HEAD(&entry->list); | 
|  |  | 
|  | resctrl_arch_rmid_idx_decode(i, &entry->closid, &entry->rmid); | 
|  | list_add_tail(&entry->list, &rmid_free_lru); | 
|  | } | 
|  |  | 
|  | /* | 
|  | * RESCTRL_RESERVED_CLOSID and RESCTRL_RESERVED_RMID are special and | 
|  | * are always allocated. These are used for the rdtgroup_default | 
|  | * control group, which will be setup later in resctrl_init(). | 
|  | */ | 
|  | idx = resctrl_arch_rmid_idx_encode(RESCTRL_RESERVED_CLOSID, | 
|  | RESCTRL_RESERVED_RMID); | 
|  | entry = __rmid_entry(idx); | 
|  | list_del(&entry->list); | 
|  |  | 
|  | out_unlock: | 
|  | mutex_unlock(&rdtgroup_mutex); | 
|  |  | 
|  | return err; | 
|  | } | 
|  |  | 
|  | static void dom_data_exit(struct rdt_resource *r) | 
|  | { | 
|  | mutex_lock(&rdtgroup_mutex); | 
|  |  | 
|  | if (!r->mon_capable) | 
|  | goto out_unlock; | 
|  |  | 
|  | if (IS_ENABLED(CONFIG_RESCTRL_RMID_DEPENDS_ON_CLOSID)) { | 
|  | kfree(closid_num_dirty_rmid); | 
|  | closid_num_dirty_rmid = NULL; | 
|  | } | 
|  |  | 
|  | kfree(rmid_ptrs); | 
|  | rmid_ptrs = NULL; | 
|  |  | 
|  | out_unlock: | 
|  | mutex_unlock(&rdtgroup_mutex); | 
|  | } | 
|  |  | 
|  | /* | 
|  | * All available events. Architecture code marks the ones that | 
|  | * are supported by a system using resctrl_enable_mon_event() | 
|  | * to set .enabled. | 
|  | */ | 
|  | struct mon_evt mon_event_all[QOS_NUM_EVENTS] = { | 
|  | [QOS_L3_OCCUP_EVENT_ID] = { | 
|  | .name	= "llc_occupancy", | 
|  | .evtid	= QOS_L3_OCCUP_EVENT_ID, | 
|  | .rid	= RDT_RESOURCE_L3, | 
|  | }, | 
|  | [QOS_L3_MBM_TOTAL_EVENT_ID] = { | 
|  | .name	= "mbm_total_bytes", | 
|  | .evtid	= QOS_L3_MBM_TOTAL_EVENT_ID, | 
|  | .rid	= RDT_RESOURCE_L3, | 
|  | }, | 
|  | [QOS_L3_MBM_LOCAL_EVENT_ID] = { | 
|  | .name	= "mbm_local_bytes", | 
|  | .evtid	= QOS_L3_MBM_LOCAL_EVENT_ID, | 
|  | .rid	= RDT_RESOURCE_L3, | 
|  | }, | 
|  | }; | 
|  |  | 
|  | void resctrl_enable_mon_event(enum resctrl_event_id eventid) | 
|  | { | 
|  | if (WARN_ON_ONCE(eventid < QOS_FIRST_EVENT || eventid >= QOS_NUM_EVENTS)) | 
|  | return; | 
|  | if (mon_event_all[eventid].enabled) { | 
|  | pr_warn("Duplicate enable for event %d\n", eventid); | 
|  | return; | 
|  | } | 
|  |  | 
|  | mon_event_all[eventid].enabled = true; | 
|  | } | 
|  |  | 
|  | bool resctrl_is_mon_event_enabled(enum resctrl_event_id eventid) | 
|  | { | 
|  | return eventid >= QOS_FIRST_EVENT && eventid < QOS_NUM_EVENTS && | 
|  | mon_event_all[eventid].enabled; | 
|  | } | 
|  |  | 
|  | u32 resctrl_get_mon_evt_cfg(enum resctrl_event_id evtid) | 
|  | { | 
|  | return mon_event_all[evtid].evt_cfg; | 
|  | } | 
|  |  | 
|  | /** | 
|  | * struct mbm_transaction - Memory transaction an MBM event can be configured with. | 
|  | * @name:	Name of memory transaction (read, write ...). | 
|  | * @val:	The bit (eg. READS_TO_LOCAL_MEM or READS_TO_REMOTE_MEM) used to | 
|  | *		represent the memory transaction within an event's configuration. | 
|  | */ | 
|  | struct mbm_transaction { | 
|  | char	name[32]; | 
|  | u32	val; | 
|  | }; | 
|  |  | 
|  | /* Decoded values for each type of memory transaction. */ | 
|  | static struct mbm_transaction mbm_transactions[NUM_MBM_TRANSACTIONS] = { | 
|  | {"local_reads", READS_TO_LOCAL_MEM}, | 
|  | {"remote_reads", READS_TO_REMOTE_MEM}, | 
|  | {"local_non_temporal_writes", NON_TEMP_WRITE_TO_LOCAL_MEM}, | 
|  | {"remote_non_temporal_writes", NON_TEMP_WRITE_TO_REMOTE_MEM}, | 
|  | {"local_reads_slow_memory", READS_TO_LOCAL_S_MEM}, | 
|  | {"remote_reads_slow_memory", READS_TO_REMOTE_S_MEM}, | 
|  | {"dirty_victim_writes_all", DIRTY_VICTIMS_TO_ALL_MEM}, | 
|  | }; | 
|  |  | 
|  | int event_filter_show(struct kernfs_open_file *of, struct seq_file *seq, void *v) | 
|  | { | 
|  | struct mon_evt *mevt = rdt_kn_parent_priv(of->kn); | 
|  | struct rdt_resource *r; | 
|  | bool sep = false; | 
|  | int ret = 0, i; | 
|  |  | 
|  | mutex_lock(&rdtgroup_mutex); | 
|  | rdt_last_cmd_clear(); | 
|  |  | 
|  | r = resctrl_arch_get_resource(mevt->rid); | 
|  | if (!resctrl_arch_mbm_cntr_assign_enabled(r)) { | 
|  | rdt_last_cmd_puts("mbm_event counter assignment mode is not enabled\n"); | 
|  | ret = -EINVAL; | 
|  | goto out_unlock; | 
|  | } | 
|  |  | 
|  | for (i = 0; i < NUM_MBM_TRANSACTIONS; i++) { | 
|  | if (mevt->evt_cfg & mbm_transactions[i].val) { | 
|  | if (sep) | 
|  | seq_putc(seq, ','); | 
|  | seq_printf(seq, "%s", mbm_transactions[i].name); | 
|  | sep = true; | 
|  | } | 
|  | } | 
|  | seq_putc(seq, '\n'); | 
|  |  | 
|  | out_unlock: | 
|  | mutex_unlock(&rdtgroup_mutex); | 
|  |  | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | int resctrl_mbm_assign_on_mkdir_show(struct kernfs_open_file *of, struct seq_file *s, | 
|  | void *v) | 
|  | { | 
|  | struct rdt_resource *r = rdt_kn_parent_priv(of->kn); | 
|  | int ret = 0; | 
|  |  | 
|  | mutex_lock(&rdtgroup_mutex); | 
|  | rdt_last_cmd_clear(); | 
|  |  | 
|  | if (!resctrl_arch_mbm_cntr_assign_enabled(r)) { | 
|  | rdt_last_cmd_puts("mbm_event counter assignment mode is not enabled\n"); | 
|  | ret = -EINVAL; | 
|  | goto out_unlock; | 
|  | } | 
|  |  | 
|  | seq_printf(s, "%u\n", r->mon.mbm_assign_on_mkdir); | 
|  |  | 
|  | out_unlock: | 
|  | mutex_unlock(&rdtgroup_mutex); | 
|  |  | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | ssize_t resctrl_mbm_assign_on_mkdir_write(struct kernfs_open_file *of, char *buf, | 
|  | size_t nbytes, loff_t off) | 
|  | { | 
|  | struct rdt_resource *r = rdt_kn_parent_priv(of->kn); | 
|  | bool value; | 
|  | int ret; | 
|  |  | 
|  | ret = kstrtobool(buf, &value); | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | mutex_lock(&rdtgroup_mutex); | 
|  | rdt_last_cmd_clear(); | 
|  |  | 
|  | if (!resctrl_arch_mbm_cntr_assign_enabled(r)) { | 
|  | rdt_last_cmd_puts("mbm_event counter assignment mode is not enabled\n"); | 
|  | ret = -EINVAL; | 
|  | goto out_unlock; | 
|  | } | 
|  |  | 
|  | r->mon.mbm_assign_on_mkdir = value; | 
|  |  | 
|  | out_unlock: | 
|  | mutex_unlock(&rdtgroup_mutex); | 
|  |  | 
|  | return ret ?: nbytes; | 
|  | } | 
|  |  | 
|  | /* | 
|  | * mbm_cntr_free_all() - Clear all the counter ID configuration details in the | 
|  | *			 domain @d. Called when mbm_assign_mode is changed. | 
|  | */ | 
|  | static void mbm_cntr_free_all(struct rdt_resource *r, struct rdt_mon_domain *d) | 
|  | { | 
|  | memset(d->cntr_cfg, 0, sizeof(*d->cntr_cfg) * r->mon.num_mbm_cntrs); | 
|  | } | 
|  |  | 
|  | /* | 
|  | * resctrl_reset_rmid_all() - Reset all non-architecture states for all the | 
|  | *			      supported RMIDs. | 
|  | */ | 
|  | static void resctrl_reset_rmid_all(struct rdt_resource *r, struct rdt_mon_domain *d) | 
|  | { | 
|  | u32 idx_limit = resctrl_arch_system_num_rmid_idx(); | 
|  | enum resctrl_event_id evt; | 
|  | int idx; | 
|  |  | 
|  | for_each_mbm_event_id(evt) { | 
|  | if (!resctrl_is_mon_event_enabled(evt)) | 
|  | continue; | 
|  | idx = MBM_STATE_IDX(evt); | 
|  | memset(d->mbm_states[idx], 0, sizeof(*d->mbm_states[0]) * idx_limit); | 
|  | } | 
|  | } | 
|  |  | 
|  | /* | 
|  | * rdtgroup_assign_cntr() - Assign/unassign the counter ID for the event, RMID | 
|  | * pair in the domain. | 
|  | * | 
|  | * Assign the counter if @assign is true else unassign the counter. Reset the | 
|  | * associated non-architectural state. | 
|  | */ | 
|  | static void rdtgroup_assign_cntr(struct rdt_resource *r, struct rdt_mon_domain *d, | 
|  | enum resctrl_event_id evtid, u32 rmid, u32 closid, | 
|  | u32 cntr_id, bool assign) | 
|  | { | 
|  | struct mbm_state *m; | 
|  |  | 
|  | resctrl_arch_config_cntr(r, d, evtid, rmid, closid, cntr_id, assign); | 
|  |  | 
|  | m = get_mbm_state(d, closid, rmid, evtid); | 
|  | if (m) | 
|  | memset(m, 0, sizeof(*m)); | 
|  | } | 
|  |  | 
|  | /* | 
|  | * rdtgroup_alloc_assign_cntr() - Allocate a counter ID and assign it to the event | 
|  | * pointed to by @mevt and the resctrl group @rdtgrp within the domain @d. | 
|  | * | 
|  | * Return: | 
|  | * 0 on success, < 0 on failure. | 
|  | */ | 
|  | static int rdtgroup_alloc_assign_cntr(struct rdt_resource *r, struct rdt_mon_domain *d, | 
|  | struct rdtgroup *rdtgrp, struct mon_evt *mevt) | 
|  | { | 
|  | int cntr_id; | 
|  |  | 
|  | /* No action required if the counter is assigned already. */ | 
|  | cntr_id = mbm_cntr_get(r, d, rdtgrp, mevt->evtid); | 
|  | if (cntr_id >= 0) | 
|  | return 0; | 
|  |  | 
|  | cntr_id = mbm_cntr_alloc(r, d, rdtgrp, mevt->evtid); | 
|  | if (cntr_id < 0) { | 
|  | rdt_last_cmd_printf("Failed to allocate counter for %s in domain %d\n", | 
|  | mevt->name, d->hdr.id); | 
|  | return cntr_id; | 
|  | } | 
|  |  | 
|  | rdtgroup_assign_cntr(r, d, mevt->evtid, rdtgrp->mon.rmid, rdtgrp->closid, cntr_id, true); | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | /* | 
|  | * rdtgroup_assign_cntr_event() - Assign a hardware counter for the event in | 
|  | * @mevt to the resctrl group @rdtgrp. Assign counters to all domains if @d is | 
|  | * NULL; otherwise, assign the counter to the specified domain @d. | 
|  | * | 
|  | * If all counters in a domain are already in use, rdtgroup_alloc_assign_cntr() | 
|  | * will fail. The assignment process will abort at the first failure encountered | 
|  | * during domain traversal, which may result in the event being only partially | 
|  | * assigned. | 
|  | * | 
|  | * Return: | 
|  | * 0 on success, < 0 on failure. | 
|  | */ | 
|  | static int rdtgroup_assign_cntr_event(struct rdt_mon_domain *d, struct rdtgroup *rdtgrp, | 
|  | struct mon_evt *mevt) | 
|  | { | 
|  | struct rdt_resource *r = resctrl_arch_get_resource(mevt->rid); | 
|  | int ret = 0; | 
|  |  | 
|  | if (!d) { | 
|  | list_for_each_entry(d, &r->mon_domains, hdr.list) { | 
|  | ret = rdtgroup_alloc_assign_cntr(r, d, rdtgrp, mevt); | 
|  | if (ret) | 
|  | return ret; | 
|  | } | 
|  | } else { | 
|  | ret = rdtgroup_alloc_assign_cntr(r, d, rdtgrp, mevt); | 
|  | } | 
|  |  | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | /* | 
|  | * rdtgroup_assign_cntrs() - Assign counters to MBM events. Called when | 
|  | *			     a new group is created. | 
|  | * | 
|  | * Each group can accommodate two counters per domain: one for the total | 
|  | * event and one for the local event. Assignments may fail due to the limited | 
|  | * number of counters. However, it is not necessary to fail the group creation | 
|  | * and thus no failure is returned. Users have the option to modify the | 
|  | * counter assignments after the group has been created. | 
|  | */ | 
|  | void rdtgroup_assign_cntrs(struct rdtgroup *rdtgrp) | 
|  | { | 
|  | struct rdt_resource *r = resctrl_arch_get_resource(RDT_RESOURCE_L3); | 
|  |  | 
|  | if (!r->mon_capable || !resctrl_arch_mbm_cntr_assign_enabled(r) || | 
|  | !r->mon.mbm_assign_on_mkdir) | 
|  | return; | 
|  |  | 
|  | if (resctrl_is_mon_event_enabled(QOS_L3_MBM_TOTAL_EVENT_ID)) | 
|  | rdtgroup_assign_cntr_event(NULL, rdtgrp, | 
|  | &mon_event_all[QOS_L3_MBM_TOTAL_EVENT_ID]); | 
|  |  | 
|  | if (resctrl_is_mon_event_enabled(QOS_L3_MBM_LOCAL_EVENT_ID)) | 
|  | rdtgroup_assign_cntr_event(NULL, rdtgrp, | 
|  | &mon_event_all[QOS_L3_MBM_LOCAL_EVENT_ID]); | 
|  | } | 
|  |  | 
|  | /* | 
|  | * rdtgroup_free_unassign_cntr() - Unassign and reset the counter ID configuration | 
|  | * for the event pointed to by @mevt within the domain @d and resctrl group @rdtgrp. | 
|  | */ | 
|  | static void rdtgroup_free_unassign_cntr(struct rdt_resource *r, struct rdt_mon_domain *d, | 
|  | struct rdtgroup *rdtgrp, struct mon_evt *mevt) | 
|  | { | 
|  | int cntr_id; | 
|  |  | 
|  | cntr_id = mbm_cntr_get(r, d, rdtgrp, mevt->evtid); | 
|  |  | 
|  | /* If there is no cntr_id assigned, nothing to do */ | 
|  | if (cntr_id < 0) | 
|  | return; | 
|  |  | 
|  | rdtgroup_assign_cntr(r, d, mevt->evtid, rdtgrp->mon.rmid, rdtgrp->closid, cntr_id, false); | 
|  |  | 
|  | mbm_cntr_free(d, cntr_id); | 
|  | } | 
|  |  | 
|  | /* | 
|  | * rdtgroup_unassign_cntr_event() - Unassign a hardware counter associated with | 
|  | * the event structure @mevt from the domain @d and the group @rdtgrp. Unassign | 
|  | * the counters from all the domains if @d is NULL else unassign from @d. | 
|  | */ | 
|  | static void rdtgroup_unassign_cntr_event(struct rdt_mon_domain *d, struct rdtgroup *rdtgrp, | 
|  | struct mon_evt *mevt) | 
|  | { | 
|  | struct rdt_resource *r = resctrl_arch_get_resource(mevt->rid); | 
|  |  | 
|  | if (!d) { | 
|  | list_for_each_entry(d, &r->mon_domains, hdr.list) | 
|  | rdtgroup_free_unassign_cntr(r, d, rdtgrp, mevt); | 
|  | } else { | 
|  | rdtgroup_free_unassign_cntr(r, d, rdtgrp, mevt); | 
|  | } | 
|  | } | 
|  |  | 
|  | /* | 
|  | * rdtgroup_unassign_cntrs() - Unassign the counters associated with MBM events. | 
|  | *			       Called when a group is deleted. | 
|  | */ | 
|  | void rdtgroup_unassign_cntrs(struct rdtgroup *rdtgrp) | 
|  | { | 
|  | struct rdt_resource *r = resctrl_arch_get_resource(RDT_RESOURCE_L3); | 
|  |  | 
|  | if (!r->mon_capable || !resctrl_arch_mbm_cntr_assign_enabled(r)) | 
|  | return; | 
|  |  | 
|  | if (resctrl_is_mon_event_enabled(QOS_L3_MBM_TOTAL_EVENT_ID)) | 
|  | rdtgroup_unassign_cntr_event(NULL, rdtgrp, | 
|  | &mon_event_all[QOS_L3_MBM_TOTAL_EVENT_ID]); | 
|  |  | 
|  | if (resctrl_is_mon_event_enabled(QOS_L3_MBM_LOCAL_EVENT_ID)) | 
|  | rdtgroup_unassign_cntr_event(NULL, rdtgrp, | 
|  | &mon_event_all[QOS_L3_MBM_LOCAL_EVENT_ID]); | 
|  | } | 
|  |  | 
|  | static int resctrl_parse_mem_transactions(char *tok, u32 *val) | 
|  | { | 
|  | u32 temp_val = 0; | 
|  | char *evt_str; | 
|  | bool found; | 
|  | int i; | 
|  |  | 
|  | next_config: | 
|  | if (!tok || tok[0] == '\0') { | 
|  | *val = temp_val; | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | /* Start processing the strings for each memory transaction type */ | 
|  | evt_str = strim(strsep(&tok, ",")); | 
|  | found = false; | 
|  | for (i = 0; i < NUM_MBM_TRANSACTIONS; i++) { | 
|  | if (!strcmp(mbm_transactions[i].name, evt_str)) { | 
|  | temp_val |= mbm_transactions[i].val; | 
|  | found = true; | 
|  | break; | 
|  | } | 
|  | } | 
|  |  | 
|  | if (!found) { | 
|  | rdt_last_cmd_printf("Invalid memory transaction type %s\n", evt_str); | 
|  | return -EINVAL; | 
|  | } | 
|  |  | 
|  | goto next_config; | 
|  | } | 
|  |  | 
|  | /* | 
|  | * rdtgroup_update_cntr_event - Update the counter assignments for the event | 
|  | *				in a group. | 
|  | * @r:		Resource to which update needs to be done. | 
|  | * @rdtgrp:	Resctrl group. | 
|  | * @evtid:	MBM monitor event. | 
|  | */ | 
|  | static void rdtgroup_update_cntr_event(struct rdt_resource *r, struct rdtgroup *rdtgrp, | 
|  | enum resctrl_event_id evtid) | 
|  | { | 
|  | struct rdt_mon_domain *d; | 
|  | int cntr_id; | 
|  |  | 
|  | list_for_each_entry(d, &r->mon_domains, hdr.list) { | 
|  | cntr_id = mbm_cntr_get(r, d, rdtgrp, evtid); | 
|  | if (cntr_id >= 0) | 
|  | rdtgroup_assign_cntr(r, d, evtid, rdtgrp->mon.rmid, | 
|  | rdtgrp->closid, cntr_id, true); | 
|  | } | 
|  | } | 
|  |  | 
|  | /* | 
|  | * resctrl_update_cntr_allrdtgrp - Update the counter assignments for the event | 
|  | *				   for all the groups. | 
|  | * @mevt	MBM Monitor event. | 
|  | */ | 
|  | static void resctrl_update_cntr_allrdtgrp(struct mon_evt *mevt) | 
|  | { | 
|  | struct rdt_resource *r = resctrl_arch_get_resource(mevt->rid); | 
|  | struct rdtgroup *prgrp, *crgrp; | 
|  |  | 
|  | /* | 
|  | * Find all the groups where the event is assigned and update the | 
|  | * configuration of existing assignments. | 
|  | */ | 
|  | list_for_each_entry(prgrp, &rdt_all_groups, rdtgroup_list) { | 
|  | rdtgroup_update_cntr_event(r, prgrp, mevt->evtid); | 
|  |  | 
|  | list_for_each_entry(crgrp, &prgrp->mon.crdtgrp_list, mon.crdtgrp_list) | 
|  | rdtgroup_update_cntr_event(r, crgrp, mevt->evtid); | 
|  | } | 
|  | } | 
|  |  | 
|  | ssize_t event_filter_write(struct kernfs_open_file *of, char *buf, size_t nbytes, | 
|  | loff_t off) | 
|  | { | 
|  | struct mon_evt *mevt = rdt_kn_parent_priv(of->kn); | 
|  | struct rdt_resource *r; | 
|  | u32 evt_cfg = 0; | 
|  | int ret = 0; | 
|  |  | 
|  | /* Valid input requires a trailing newline */ | 
|  | if (nbytes == 0 || buf[nbytes - 1] != '\n') | 
|  | return -EINVAL; | 
|  |  | 
|  | buf[nbytes - 1] = '\0'; | 
|  |  | 
|  | cpus_read_lock(); | 
|  | mutex_lock(&rdtgroup_mutex); | 
|  |  | 
|  | rdt_last_cmd_clear(); | 
|  |  | 
|  | r = resctrl_arch_get_resource(mevt->rid); | 
|  | if (!resctrl_arch_mbm_cntr_assign_enabled(r)) { | 
|  | rdt_last_cmd_puts("mbm_event counter assignment mode is not enabled\n"); | 
|  | ret = -EINVAL; | 
|  | goto out_unlock; | 
|  | } | 
|  |  | 
|  | ret = resctrl_parse_mem_transactions(buf, &evt_cfg); | 
|  | if (!ret && mevt->evt_cfg != evt_cfg) { | 
|  | mevt->evt_cfg = evt_cfg; | 
|  | resctrl_update_cntr_allrdtgrp(mevt); | 
|  | } | 
|  |  | 
|  | out_unlock: | 
|  | mutex_unlock(&rdtgroup_mutex); | 
|  | cpus_read_unlock(); | 
|  |  | 
|  | return ret ?: nbytes; | 
|  | } | 
|  |  | 
|  | int resctrl_mbm_assign_mode_show(struct kernfs_open_file *of, | 
|  | struct seq_file *s, void *v) | 
|  | { | 
|  | struct rdt_resource *r = rdt_kn_parent_priv(of->kn); | 
|  | bool enabled; | 
|  |  | 
|  | mutex_lock(&rdtgroup_mutex); | 
|  | enabled = resctrl_arch_mbm_cntr_assign_enabled(r); | 
|  |  | 
|  | if (r->mon.mbm_cntr_assignable) { | 
|  | if (enabled) | 
|  | seq_puts(s, "[mbm_event]\n"); | 
|  | else | 
|  | seq_puts(s, "[default]\n"); | 
|  |  | 
|  | if (!IS_ENABLED(CONFIG_RESCTRL_ASSIGN_FIXED)) { | 
|  | if (enabled) | 
|  | seq_puts(s, "default\n"); | 
|  | else | 
|  | seq_puts(s, "mbm_event\n"); | 
|  | } | 
|  | } else { | 
|  | seq_puts(s, "[default]\n"); | 
|  | } | 
|  |  | 
|  | mutex_unlock(&rdtgroup_mutex); | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | ssize_t resctrl_mbm_assign_mode_write(struct kernfs_open_file *of, char *buf, | 
|  | size_t nbytes, loff_t off) | 
|  | { | 
|  | struct rdt_resource *r = rdt_kn_parent_priv(of->kn); | 
|  | struct rdt_mon_domain *d; | 
|  | int ret = 0; | 
|  | bool enable; | 
|  |  | 
|  | /* Valid input requires a trailing newline */ | 
|  | if (nbytes == 0 || buf[nbytes - 1] != '\n') | 
|  | return -EINVAL; | 
|  |  | 
|  | buf[nbytes - 1] = '\0'; | 
|  |  | 
|  | cpus_read_lock(); | 
|  | mutex_lock(&rdtgroup_mutex); | 
|  |  | 
|  | rdt_last_cmd_clear(); | 
|  |  | 
|  | if (!strcmp(buf, "default")) { | 
|  | enable = 0; | 
|  | } else if (!strcmp(buf, "mbm_event")) { | 
|  | if (r->mon.mbm_cntr_assignable) { | 
|  | enable = 1; | 
|  | } else { | 
|  | ret = -EINVAL; | 
|  | rdt_last_cmd_puts("mbm_event mode is not supported\n"); | 
|  | goto out_unlock; | 
|  | } | 
|  | } else { | 
|  | ret = -EINVAL; | 
|  | rdt_last_cmd_puts("Unsupported assign mode\n"); | 
|  | goto out_unlock; | 
|  | } | 
|  |  | 
|  | if (enable != resctrl_arch_mbm_cntr_assign_enabled(r)) { | 
|  | ret = resctrl_arch_mbm_cntr_assign_set(r, enable); | 
|  | if (ret) | 
|  | goto out_unlock; | 
|  |  | 
|  | /* Update the visibility of BMEC related files */ | 
|  | resctrl_bmec_files_show(r, NULL, !enable); | 
|  |  | 
|  | /* | 
|  | * Initialize the default memory transaction values for | 
|  | * total and local events. | 
|  | */ | 
|  | if (resctrl_is_mon_event_enabled(QOS_L3_MBM_TOTAL_EVENT_ID)) | 
|  | mon_event_all[QOS_L3_MBM_TOTAL_EVENT_ID].evt_cfg = r->mon.mbm_cfg_mask; | 
|  | if (resctrl_is_mon_event_enabled(QOS_L3_MBM_LOCAL_EVENT_ID)) | 
|  | mon_event_all[QOS_L3_MBM_LOCAL_EVENT_ID].evt_cfg = r->mon.mbm_cfg_mask & | 
|  | (READS_TO_LOCAL_MEM | | 
|  | READS_TO_LOCAL_S_MEM | | 
|  | NON_TEMP_WRITE_TO_LOCAL_MEM); | 
|  | /* Enable auto assignment when switching to "mbm_event" mode */ | 
|  | if (enable) | 
|  | r->mon.mbm_assign_on_mkdir = true; | 
|  | /* | 
|  | * Reset all the non-achitectural RMID state and assignable counters. | 
|  | */ | 
|  | list_for_each_entry(d, &r->mon_domains, hdr.list) { | 
|  | mbm_cntr_free_all(r, d); | 
|  | resctrl_reset_rmid_all(r, d); | 
|  | } | 
|  | } | 
|  |  | 
|  | out_unlock: | 
|  | mutex_unlock(&rdtgroup_mutex); | 
|  | cpus_read_unlock(); | 
|  |  | 
|  | return ret ?: nbytes; | 
|  | } | 
|  |  | 
|  | int resctrl_num_mbm_cntrs_show(struct kernfs_open_file *of, | 
|  | struct seq_file *s, void *v) | 
|  | { | 
|  | struct rdt_resource *r = rdt_kn_parent_priv(of->kn); | 
|  | struct rdt_mon_domain *dom; | 
|  | bool sep = false; | 
|  |  | 
|  | cpus_read_lock(); | 
|  | mutex_lock(&rdtgroup_mutex); | 
|  |  | 
|  | list_for_each_entry(dom, &r->mon_domains, hdr.list) { | 
|  | if (sep) | 
|  | seq_putc(s, ';'); | 
|  |  | 
|  | seq_printf(s, "%d=%d", dom->hdr.id, r->mon.num_mbm_cntrs); | 
|  | sep = true; | 
|  | } | 
|  | seq_putc(s, '\n'); | 
|  |  | 
|  | mutex_unlock(&rdtgroup_mutex); | 
|  | cpus_read_unlock(); | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | int resctrl_available_mbm_cntrs_show(struct kernfs_open_file *of, | 
|  | struct seq_file *s, void *v) | 
|  | { | 
|  | struct rdt_resource *r = rdt_kn_parent_priv(of->kn); | 
|  | struct rdt_mon_domain *dom; | 
|  | bool sep = false; | 
|  | u32 cntrs, i; | 
|  | int ret = 0; | 
|  |  | 
|  | cpus_read_lock(); | 
|  | mutex_lock(&rdtgroup_mutex); | 
|  |  | 
|  | rdt_last_cmd_clear(); | 
|  |  | 
|  | if (!resctrl_arch_mbm_cntr_assign_enabled(r)) { | 
|  | rdt_last_cmd_puts("mbm_event counter assignment mode is not enabled\n"); | 
|  | ret = -EINVAL; | 
|  | goto out_unlock; | 
|  | } | 
|  |  | 
|  | list_for_each_entry(dom, &r->mon_domains, hdr.list) { | 
|  | if (sep) | 
|  | seq_putc(s, ';'); | 
|  |  | 
|  | cntrs = 0; | 
|  | for (i = 0; i < r->mon.num_mbm_cntrs; i++) { | 
|  | if (!dom->cntr_cfg[i].rdtgrp) | 
|  | cntrs++; | 
|  | } | 
|  |  | 
|  | seq_printf(s, "%d=%u", dom->hdr.id, cntrs); | 
|  | sep = true; | 
|  | } | 
|  | seq_putc(s, '\n'); | 
|  |  | 
|  | out_unlock: | 
|  | mutex_unlock(&rdtgroup_mutex); | 
|  | cpus_read_unlock(); | 
|  |  | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | int mbm_L3_assignments_show(struct kernfs_open_file *of, struct seq_file *s, void *v) | 
|  | { | 
|  | struct rdt_resource *r = resctrl_arch_get_resource(RDT_RESOURCE_L3); | 
|  | struct rdt_mon_domain *d; | 
|  | struct rdtgroup *rdtgrp; | 
|  | struct mon_evt *mevt; | 
|  | int ret = 0; | 
|  | bool sep; | 
|  |  | 
|  | rdtgrp = rdtgroup_kn_lock_live(of->kn); | 
|  | if (!rdtgrp) { | 
|  | ret = -ENOENT; | 
|  | goto out_unlock; | 
|  | } | 
|  |  | 
|  | rdt_last_cmd_clear(); | 
|  | if (!resctrl_arch_mbm_cntr_assign_enabled(r)) { | 
|  | rdt_last_cmd_puts("mbm_event counter assignment mode is not enabled\n"); | 
|  | ret = -EINVAL; | 
|  | goto out_unlock; | 
|  | } | 
|  |  | 
|  | for_each_mon_event(mevt) { | 
|  | if (mevt->rid != r->rid || !mevt->enabled || !resctrl_is_mbm_event(mevt->evtid)) | 
|  | continue; | 
|  |  | 
|  | sep = false; | 
|  | seq_printf(s, "%s:", mevt->name); | 
|  | list_for_each_entry(d, &r->mon_domains, hdr.list) { | 
|  | if (sep) | 
|  | seq_putc(s, ';'); | 
|  |  | 
|  | if (mbm_cntr_get(r, d, rdtgrp, mevt->evtid) < 0) | 
|  | seq_printf(s, "%d=_", d->hdr.id); | 
|  | else | 
|  | seq_printf(s, "%d=e", d->hdr.id); | 
|  |  | 
|  | sep = true; | 
|  | } | 
|  | seq_putc(s, '\n'); | 
|  | } | 
|  |  | 
|  | out_unlock: | 
|  | rdtgroup_kn_unlock(of->kn); | 
|  |  | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | /* | 
|  | * mbm_get_mon_event_by_name() - Return the mon_evt entry for the matching | 
|  | * event name. | 
|  | */ | 
|  | static struct mon_evt *mbm_get_mon_event_by_name(struct rdt_resource *r, char *name) | 
|  | { | 
|  | struct mon_evt *mevt; | 
|  |  | 
|  | for_each_mon_event(mevt) { | 
|  | if (mevt->rid == r->rid && mevt->enabled && | 
|  | resctrl_is_mbm_event(mevt->evtid) && | 
|  | !strcmp(mevt->name, name)) | 
|  | return mevt; | 
|  | } | 
|  |  | 
|  | return NULL; | 
|  | } | 
|  |  | 
|  | static int rdtgroup_modify_assign_state(char *assign, struct rdt_mon_domain *d, | 
|  | struct rdtgroup *rdtgrp, struct mon_evt *mevt) | 
|  | { | 
|  | int ret = 0; | 
|  |  | 
|  | if (!assign || strlen(assign) != 1) | 
|  | return -EINVAL; | 
|  |  | 
|  | switch (*assign) { | 
|  | case 'e': | 
|  | ret = rdtgroup_assign_cntr_event(d, rdtgrp, mevt); | 
|  | break; | 
|  | case '_': | 
|  | rdtgroup_unassign_cntr_event(d, rdtgrp, mevt); | 
|  | break; | 
|  | default: | 
|  | ret = -EINVAL; | 
|  | break; | 
|  | } | 
|  |  | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | static int resctrl_parse_mbm_assignment(struct rdt_resource *r, struct rdtgroup *rdtgrp, | 
|  | char *event, char *tok) | 
|  | { | 
|  | struct rdt_mon_domain *d; | 
|  | unsigned long dom_id = 0; | 
|  | char *dom_str, *id_str; | 
|  | struct mon_evt *mevt; | 
|  | int ret; | 
|  |  | 
|  | mevt = mbm_get_mon_event_by_name(r, event); | 
|  | if (!mevt) { | 
|  | rdt_last_cmd_printf("Invalid event %s\n", event); | 
|  | return -ENOENT; | 
|  | } | 
|  |  | 
|  | next: | 
|  | if (!tok || tok[0] == '\0') | 
|  | return 0; | 
|  |  | 
|  | /* Start processing the strings for each domain */ | 
|  | dom_str = strim(strsep(&tok, ";")); | 
|  |  | 
|  | id_str = strsep(&dom_str, "="); | 
|  |  | 
|  | /* Check for domain id '*' which means all domains */ | 
|  | if (id_str && *id_str == '*') { | 
|  | ret = rdtgroup_modify_assign_state(dom_str, NULL, rdtgrp, mevt); | 
|  | if (ret) | 
|  | rdt_last_cmd_printf("Assign operation '%s:*=%s' failed\n", | 
|  | event, dom_str); | 
|  | return ret; | 
|  | } else if (!id_str || kstrtoul(id_str, 10, &dom_id)) { | 
|  | rdt_last_cmd_puts("Missing domain id\n"); | 
|  | return -EINVAL; | 
|  | } | 
|  |  | 
|  | /* Verify if the dom_id is valid */ | 
|  | list_for_each_entry(d, &r->mon_domains, hdr.list) { | 
|  | if (d->hdr.id == dom_id) { | 
|  | ret = rdtgroup_modify_assign_state(dom_str, d, rdtgrp, mevt); | 
|  | if (ret) { | 
|  | rdt_last_cmd_printf("Assign operation '%s:%ld=%s' failed\n", | 
|  | event, dom_id, dom_str); | 
|  | return ret; | 
|  | } | 
|  | goto next; | 
|  | } | 
|  | } | 
|  |  | 
|  | rdt_last_cmd_printf("Invalid domain id %ld\n", dom_id); | 
|  | return -EINVAL; | 
|  | } | 
|  |  | 
|  | ssize_t mbm_L3_assignments_write(struct kernfs_open_file *of, char *buf, | 
|  | size_t nbytes, loff_t off) | 
|  | { | 
|  | struct rdt_resource *r = resctrl_arch_get_resource(RDT_RESOURCE_L3); | 
|  | struct rdtgroup *rdtgrp; | 
|  | char *token, *event; | 
|  | int ret = 0; | 
|  |  | 
|  | /* Valid input requires a trailing newline */ | 
|  | if (nbytes == 0 || buf[nbytes - 1] != '\n') | 
|  | return -EINVAL; | 
|  |  | 
|  | buf[nbytes - 1] = '\0'; | 
|  |  | 
|  | rdtgrp = rdtgroup_kn_lock_live(of->kn); | 
|  | if (!rdtgrp) { | 
|  | rdtgroup_kn_unlock(of->kn); | 
|  | return -ENOENT; | 
|  | } | 
|  | rdt_last_cmd_clear(); | 
|  |  | 
|  | if (!resctrl_arch_mbm_cntr_assign_enabled(r)) { | 
|  | rdt_last_cmd_puts("mbm_event mode is not enabled\n"); | 
|  | rdtgroup_kn_unlock(of->kn); | 
|  | return -EINVAL; | 
|  | } | 
|  |  | 
|  | while ((token = strsep(&buf, "\n")) != NULL) { | 
|  | /* | 
|  | * The write command follows the following format: | 
|  | * "<Event>:<Domain ID>=<Assignment state>" | 
|  | * Extract the event name first. | 
|  | */ | 
|  | event = strsep(&token, ":"); | 
|  |  | 
|  | ret = resctrl_parse_mbm_assignment(r, rdtgrp, event, token); | 
|  | if (ret) | 
|  | break; | 
|  | } | 
|  |  | 
|  | rdtgroup_kn_unlock(of->kn); | 
|  |  | 
|  | return ret ?: nbytes; | 
|  | } | 
|  |  | 
|  | /** | 
|  | * resctrl_mon_resource_init() - Initialise global monitoring structures. | 
|  | * | 
|  | * Allocate and initialise global monitor resources that do not belong to a | 
|  | * specific domain. i.e. the rmid_ptrs[] used for the limbo and free lists. | 
|  | * Called once during boot after the struct rdt_resource's have been configured | 
|  | * but before the filesystem is mounted. | 
|  | * Resctrl's cpuhp callbacks may be called before this point to bring a domain | 
|  | * online. | 
|  | * | 
|  | * Returns 0 for success, or -ENOMEM. | 
|  | */ | 
|  | int resctrl_mon_resource_init(void) | 
|  | { | 
|  | struct rdt_resource *r = resctrl_arch_get_resource(RDT_RESOURCE_L3); | 
|  | int ret; | 
|  |  | 
|  | if (!r->mon_capable) | 
|  | return 0; | 
|  |  | 
|  | ret = dom_data_init(r); | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | if (resctrl_arch_is_evt_configurable(QOS_L3_MBM_TOTAL_EVENT_ID)) { | 
|  | mon_event_all[QOS_L3_MBM_TOTAL_EVENT_ID].configurable = true; | 
|  | resctrl_file_fflags_init("mbm_total_bytes_config", | 
|  | RFTYPE_MON_INFO | RFTYPE_RES_CACHE); | 
|  | } | 
|  | if (resctrl_arch_is_evt_configurable(QOS_L3_MBM_LOCAL_EVENT_ID)) { | 
|  | mon_event_all[QOS_L3_MBM_LOCAL_EVENT_ID].configurable = true; | 
|  | resctrl_file_fflags_init("mbm_local_bytes_config", | 
|  | RFTYPE_MON_INFO | RFTYPE_RES_CACHE); | 
|  | } | 
|  |  | 
|  | if (resctrl_is_mon_event_enabled(QOS_L3_MBM_LOCAL_EVENT_ID)) | 
|  | mba_mbps_default_event = QOS_L3_MBM_LOCAL_EVENT_ID; | 
|  | else if (resctrl_is_mon_event_enabled(QOS_L3_MBM_TOTAL_EVENT_ID)) | 
|  | mba_mbps_default_event = QOS_L3_MBM_TOTAL_EVENT_ID; | 
|  |  | 
|  | if (r->mon.mbm_cntr_assignable) { | 
|  | if (!resctrl_is_mon_event_enabled(QOS_L3_MBM_TOTAL_EVENT_ID)) | 
|  | resctrl_enable_mon_event(QOS_L3_MBM_TOTAL_EVENT_ID); | 
|  | if (!resctrl_is_mon_event_enabled(QOS_L3_MBM_LOCAL_EVENT_ID)) | 
|  | resctrl_enable_mon_event(QOS_L3_MBM_LOCAL_EVENT_ID); | 
|  | mon_event_all[QOS_L3_MBM_TOTAL_EVENT_ID].evt_cfg = r->mon.mbm_cfg_mask; | 
|  | mon_event_all[QOS_L3_MBM_LOCAL_EVENT_ID].evt_cfg = r->mon.mbm_cfg_mask & | 
|  | (READS_TO_LOCAL_MEM | | 
|  | READS_TO_LOCAL_S_MEM | | 
|  | NON_TEMP_WRITE_TO_LOCAL_MEM); | 
|  | r->mon.mbm_assign_on_mkdir = true; | 
|  | resctrl_file_fflags_init("num_mbm_cntrs", | 
|  | RFTYPE_MON_INFO | RFTYPE_RES_CACHE); | 
|  | resctrl_file_fflags_init("available_mbm_cntrs", | 
|  | RFTYPE_MON_INFO | RFTYPE_RES_CACHE); | 
|  | resctrl_file_fflags_init("event_filter", RFTYPE_ASSIGN_CONFIG); | 
|  | resctrl_file_fflags_init("mbm_assign_on_mkdir", RFTYPE_MON_INFO | | 
|  | RFTYPE_RES_CACHE); | 
|  | resctrl_file_fflags_init("mbm_L3_assignments", RFTYPE_MON_BASE); | 
|  | } | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | void resctrl_mon_resource_exit(void) | 
|  | { | 
|  | struct rdt_resource *r = resctrl_arch_get_resource(RDT_RESOURCE_L3); | 
|  |  | 
|  | dom_data_exit(r); | 
|  | } |