1 | // SPDX-License-Identifier: GPL-2.0-only |
2 | /* |
3 | * MIPS cacheinfo support |
4 | */ |
5 | #include <linux/cacheinfo.h> |
6 | |
7 | /* Populates leaf and increments to next leaf */ |
8 | #define populate_cache(cache, leaf, c_level, c_type) \ |
9 | do { \ |
10 | leaf->type = c_type; \ |
11 | leaf->level = c_level; \ |
12 | leaf->coherency_line_size = c->cache.linesz; \ |
13 | leaf->number_of_sets = c->cache.sets; \ |
14 | leaf->ways_of_associativity = c->cache.ways; \ |
15 | leaf->size = c->cache.linesz * c->cache.sets * \ |
16 | c->cache.ways; \ |
17 | leaf++; \ |
18 | } while (0) |
19 | |
20 | int init_cache_level(unsigned int cpu) |
21 | { |
22 | struct cpuinfo_mips *c = ¤t_cpu_data; |
23 | struct cpu_cacheinfo *this_cpu_ci = get_cpu_cacheinfo(cpu); |
24 | int levels = 0, leaves = 0; |
25 | |
26 | /* |
27 | * If Dcache is not set, we assume the cache structures |
28 | * are not properly initialized. |
29 | */ |
30 | if (c->dcache.waysize) |
31 | levels += 1; |
32 | else |
33 | return -ENOENT; |
34 | |
35 | |
36 | leaves += (c->icache.waysize) ? 2 : 1; |
37 | |
38 | if (c->vcache.waysize) { |
39 | levels++; |
40 | leaves++; |
41 | } |
42 | |
43 | if (c->scache.waysize) { |
44 | levels++; |
45 | leaves++; |
46 | } |
47 | |
48 | if (c->tcache.waysize) { |
49 | levels++; |
50 | leaves++; |
51 | } |
52 | |
53 | this_cpu_ci->num_levels = levels; |
54 | this_cpu_ci->num_leaves = leaves; |
55 | return 0; |
56 | } |
57 | |
58 | static void fill_cpumask_siblings(int cpu, cpumask_t *cpu_map) |
59 | { |
60 | int cpu1; |
61 | |
62 | for_each_possible_cpu(cpu1) |
63 | if (cpus_are_siblings(cpu, cpu1)) |
64 | cpumask_set_cpu(cpu: cpu1, dstp: cpu_map); |
65 | } |
66 | |
67 | static void fill_cpumask_cluster(int cpu, cpumask_t *cpu_map) |
68 | { |
69 | int cpu1; |
70 | int cluster = cpu_cluster(&cpu_data[cpu]); |
71 | |
72 | for_each_possible_cpu(cpu1) |
73 | if (cpu_cluster(&cpu_data[cpu1]) == cluster) |
74 | cpumask_set_cpu(cpu: cpu1, dstp: cpu_map); |
75 | } |
76 | |
77 | int populate_cache_leaves(unsigned int cpu) |
78 | { |
79 | struct cpuinfo_mips *c = ¤t_cpu_data; |
80 | struct cpu_cacheinfo *this_cpu_ci = get_cpu_cacheinfo(cpu); |
81 | struct cacheinfo *this_leaf = this_cpu_ci->info_list; |
82 | int level = 1; |
83 | |
84 | if (c->icache.waysize) { |
85 | /* I/D caches are per core */ |
86 | fill_cpumask_siblings(cpu, cpu_map: &this_leaf->shared_cpu_map); |
87 | populate_cache(dcache, this_leaf, level, CACHE_TYPE_DATA); |
88 | fill_cpumask_siblings(cpu, cpu_map: &this_leaf->shared_cpu_map); |
89 | populate_cache(icache, this_leaf, level, CACHE_TYPE_INST); |
90 | level++; |
91 | } else { |
92 | populate_cache(dcache, this_leaf, level, CACHE_TYPE_UNIFIED); |
93 | level++; |
94 | } |
95 | |
96 | if (c->vcache.waysize) { |
97 | /* Vcache is per core as well */ |
98 | fill_cpumask_siblings(cpu, cpu_map: &this_leaf->shared_cpu_map); |
99 | populate_cache(vcache, this_leaf, level, CACHE_TYPE_UNIFIED); |
100 | level++; |
101 | } |
102 | |
103 | if (c->scache.waysize) { |
104 | /* Scache is per cluster */ |
105 | fill_cpumask_cluster(cpu, cpu_map: &this_leaf->shared_cpu_map); |
106 | populate_cache(scache, this_leaf, level, CACHE_TYPE_UNIFIED); |
107 | level++; |
108 | } |
109 | |
110 | if (c->tcache.waysize) |
111 | populate_cache(tcache, this_leaf, level, CACHE_TYPE_UNIFIED); |
112 | |
113 | this_cpu_ci->cpu_map_populated = true; |
114 | |
115 | return 0; |
116 | } |
117 | |