|  | // SPDX-License-Identifier: GPL-2.0 | 
|  |  | 
|  | #include <linux/version.h> | 
|  | #include <linux/ptrace.h> | 
|  | #include <uapi/linux/bpf.h> | 
|  | #include <bpf/bpf_helpers.h> | 
|  |  | 
|  | /* | 
|  | * The CPU number, cstate number and pstate number are based | 
|  | * on 96boards Hikey with octa CA53 CPUs. | 
|  | * | 
|  | * Every CPU have three idle states for cstate: | 
|  | *   WFI, CPU_OFF, CLUSTER_OFF | 
|  | * | 
|  | * Every CPU have 5 operating points: | 
|  | *   208MHz, 432MHz, 729MHz, 960MHz, 1200MHz | 
|  | * | 
|  | * This code is based on these assumption and other platforms | 
|  | * need to adjust these definitions. | 
|  | */ | 
|  | #define MAX_CPU			8 | 
|  | #define MAX_PSTATE_ENTRIES	5 | 
|  | #define MAX_CSTATE_ENTRIES	3 | 
|  |  | 
|  | static int cpu_opps[] = { 208000, 432000, 729000, 960000, 1200000 }; | 
|  |  | 
|  | /* | 
|  | * my_map structure is used to record cstate and pstate index and | 
|  | * timestamp (Idx, Ts), when new event incoming we need to update | 
|  | * combination for new state index and timestamp (Idx`, Ts`). | 
|  | * | 
|  | * Based on (Idx, Ts) and (Idx`, Ts`) we can calculate the time | 
|  | * interval for the previous state: Duration(Idx) = Ts` - Ts. | 
|  | * | 
|  | * Every CPU has one below array for recording state index and | 
|  | * timestamp, and record for cstate and pstate saperately: | 
|  | * | 
|  | * +--------------------------+ | 
|  | * | cstate timestamp         | | 
|  | * +--------------------------+ | 
|  | * | cstate index             | | 
|  | * +--------------------------+ | 
|  | * | pstate timestamp         | | 
|  | * +--------------------------+ | 
|  | * | pstate index             | | 
|  | * +--------------------------+ | 
|  | */ | 
|  | #define MAP_OFF_CSTATE_TIME	0 | 
|  | #define MAP_OFF_CSTATE_IDX	1 | 
|  | #define MAP_OFF_PSTATE_TIME	2 | 
|  | #define MAP_OFF_PSTATE_IDX	3 | 
|  | #define MAP_OFF_NUM		4 | 
|  |  | 
|  | struct { | 
|  | __uint(type, BPF_MAP_TYPE_ARRAY); | 
|  | __type(key, u32); | 
|  | __type(value, u64); | 
|  | __uint(max_entries, MAX_CPU * MAP_OFF_NUM); | 
|  | } my_map SEC(".maps"); | 
|  |  | 
|  | /* cstate_duration records duration time for every idle state per CPU */ | 
|  | struct { | 
|  | __uint(type, BPF_MAP_TYPE_ARRAY); | 
|  | __type(key, u32); | 
|  | __type(value, u64); | 
|  | __uint(max_entries, MAX_CPU * MAX_CSTATE_ENTRIES); | 
|  | } cstate_duration SEC(".maps"); | 
|  |  | 
|  | /* pstate_duration records duration time for every operating point per CPU */ | 
|  | struct { | 
|  | __uint(type, BPF_MAP_TYPE_ARRAY); | 
|  | __type(key, u32); | 
|  | __type(value, u64); | 
|  | __uint(max_entries, MAX_CPU * MAX_PSTATE_ENTRIES); | 
|  | } pstate_duration SEC(".maps"); | 
|  |  | 
|  | /* | 
|  | * The trace events for cpu_idle and cpu_frequency are taken from: | 
|  | * /sys/kernel/tracing/events/power/cpu_idle/format | 
|  | * /sys/kernel/tracing/events/power/cpu_frequency/format | 
|  | * | 
|  | * These two events have same format, so define one common structure. | 
|  | */ | 
|  | struct cpu_args { | 
|  | u64 pad; | 
|  | u32 state; | 
|  | u32 cpu_id; | 
|  | }; | 
|  |  | 
|  | /* calculate pstate index, returns MAX_PSTATE_ENTRIES for failure */ | 
|  | static u32 find_cpu_pstate_idx(u32 frequency) | 
|  | { | 
|  | u32 i; | 
|  |  | 
|  | for (i = 0; i < sizeof(cpu_opps) / sizeof(u32); i++) { | 
|  | if (frequency == cpu_opps[i]) | 
|  | return i; | 
|  | } | 
|  |  | 
|  | return i; | 
|  | } | 
|  |  | 
|  | SEC("tracepoint/power/cpu_idle") | 
|  | int bpf_prog1(struct cpu_args *ctx) | 
|  | { | 
|  | u64 *cts, *pts, *cstate, *pstate, prev_state, cur_ts, delta; | 
|  | u32 key, cpu, pstate_idx; | 
|  | u64 *val; | 
|  |  | 
|  | if (ctx->cpu_id > MAX_CPU) | 
|  | return 0; | 
|  |  | 
|  | cpu = ctx->cpu_id; | 
|  |  | 
|  | key = cpu * MAP_OFF_NUM + MAP_OFF_CSTATE_TIME; | 
|  | cts = bpf_map_lookup_elem(&my_map, &key); | 
|  | if (!cts) | 
|  | return 0; | 
|  |  | 
|  | key = cpu * MAP_OFF_NUM + MAP_OFF_CSTATE_IDX; | 
|  | cstate = bpf_map_lookup_elem(&my_map, &key); | 
|  | if (!cstate) | 
|  | return 0; | 
|  |  | 
|  | key = cpu * MAP_OFF_NUM + MAP_OFF_PSTATE_TIME; | 
|  | pts = bpf_map_lookup_elem(&my_map, &key); | 
|  | if (!pts) | 
|  | return 0; | 
|  |  | 
|  | key = cpu * MAP_OFF_NUM + MAP_OFF_PSTATE_IDX; | 
|  | pstate = bpf_map_lookup_elem(&my_map, &key); | 
|  | if (!pstate) | 
|  | return 0; | 
|  |  | 
|  | prev_state = *cstate; | 
|  | *cstate = ctx->state; | 
|  |  | 
|  | if (!*cts) { | 
|  | *cts = bpf_ktime_get_ns(); | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | cur_ts = bpf_ktime_get_ns(); | 
|  | delta = cur_ts - *cts; | 
|  | *cts = cur_ts; | 
|  |  | 
|  | /* | 
|  | * When state doesn't equal to (u32)-1, the cpu will enter | 
|  | * one idle state; for this case we need to record interval | 
|  | * for the pstate. | 
|  | * | 
|  | *                 OPP2 | 
|  | *            +---------------------+ | 
|  | *     OPP1   |                     | | 
|  | *   ---------+                     | | 
|  | *                                  |  Idle state | 
|  | *                                  +--------------- | 
|  | * | 
|  | *            |<- pstate duration ->| | 
|  | *            ^                     ^ | 
|  | *           pts                  cur_ts | 
|  | */ | 
|  | if (ctx->state != (u32)-1) { | 
|  |  | 
|  | /* record pstate after have first cpu_frequency event */ | 
|  | if (!*pts) | 
|  | return 0; | 
|  |  | 
|  | delta = cur_ts - *pts; | 
|  |  | 
|  | pstate_idx = find_cpu_pstate_idx(*pstate); | 
|  | if (pstate_idx >= MAX_PSTATE_ENTRIES) | 
|  | return 0; | 
|  |  | 
|  | key = cpu * MAX_PSTATE_ENTRIES + pstate_idx; | 
|  | val = bpf_map_lookup_elem(&pstate_duration, &key); | 
|  | if (val) | 
|  | __sync_fetch_and_add((long *)val, delta); | 
|  |  | 
|  | /* | 
|  | * When state equal to (u32)-1, the cpu just exits from one | 
|  | * specific idle state; for this case we need to record | 
|  | * interval for the pstate. | 
|  | * | 
|  | *       OPP2 | 
|  | *   -----------+ | 
|  | *              |                          OPP1 | 
|  | *              |                     +----------- | 
|  | *              |     Idle state      | | 
|  | *              +---------------------+ | 
|  | * | 
|  | *              |<- cstate duration ->| | 
|  | *              ^                     ^ | 
|  | *             cts                  cur_ts | 
|  | */ | 
|  | } else { | 
|  |  | 
|  | key = cpu * MAX_CSTATE_ENTRIES + prev_state; | 
|  | val = bpf_map_lookup_elem(&cstate_duration, &key); | 
|  | if (val) | 
|  | __sync_fetch_and_add((long *)val, delta); | 
|  | } | 
|  |  | 
|  | /* Update timestamp for pstate as new start time */ | 
|  | if (*pts) | 
|  | *pts = cur_ts; | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | SEC("tracepoint/power/cpu_frequency") | 
|  | int bpf_prog2(struct cpu_args *ctx) | 
|  | { | 
|  | u64 *pts, *cstate, *pstate, cur_ts, delta; | 
|  | u32 key, cpu, pstate_idx; | 
|  | u64 *val; | 
|  |  | 
|  | cpu = ctx->cpu_id; | 
|  |  | 
|  | key = cpu * MAP_OFF_NUM + MAP_OFF_PSTATE_TIME; | 
|  | pts = bpf_map_lookup_elem(&my_map, &key); | 
|  | if (!pts) | 
|  | return 0; | 
|  |  | 
|  | key = cpu * MAP_OFF_NUM + MAP_OFF_PSTATE_IDX; | 
|  | pstate = bpf_map_lookup_elem(&my_map, &key); | 
|  | if (!pstate) | 
|  | return 0; | 
|  |  | 
|  | key = cpu * MAP_OFF_NUM + MAP_OFF_CSTATE_IDX; | 
|  | cstate = bpf_map_lookup_elem(&my_map, &key); | 
|  | if (!cstate) | 
|  | return 0; | 
|  |  | 
|  | *pstate = ctx->state; | 
|  |  | 
|  | if (!*pts) { | 
|  | *pts = bpf_ktime_get_ns(); | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | cur_ts = bpf_ktime_get_ns(); | 
|  | delta = cur_ts - *pts; | 
|  | *pts = cur_ts; | 
|  |  | 
|  | /* When CPU is in idle, bail out to skip pstate statistics */ | 
|  | if (*cstate != (u32)(-1)) | 
|  | return 0; | 
|  |  | 
|  | /* | 
|  | * The cpu changes to another different OPP (in below diagram | 
|  | * change frequency from OPP3 to OPP1), need recording interval | 
|  | * for previous frequency OPP3 and update timestamp as start | 
|  | * time for new frequency OPP1. | 
|  | * | 
|  | *                 OPP3 | 
|  | *            +---------------------+ | 
|  | *     OPP2   |                     | | 
|  | *   ---------+                     | | 
|  | *                                  |    OPP1 | 
|  | *                                  +--------------- | 
|  | * | 
|  | *            |<- pstate duration ->| | 
|  | *            ^                     ^ | 
|  | *           pts                  cur_ts | 
|  | */ | 
|  | pstate_idx = find_cpu_pstate_idx(*pstate); | 
|  | if (pstate_idx >= MAX_PSTATE_ENTRIES) | 
|  | return 0; | 
|  |  | 
|  | key = cpu * MAX_PSTATE_ENTRIES + pstate_idx; | 
|  | val = bpf_map_lookup_elem(&pstate_duration, &key); | 
|  | if (val) | 
|  | __sync_fetch_and_add((long *)val, delta); | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | char _license[] SEC("license") = "GPL"; | 
|  | u32 _version SEC("version") = LINUX_VERSION_CODE; |