123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116 |
- // SPDX-License-Identifier: GPL-2.0-only
- /*
- * MIPS cacheinfo support
- */
- #include <linux/cacheinfo.h>
- /* Populates leaf and increments to next leaf */
- #define populate_cache(cache, leaf, c_level, c_type) \
- do { \
- leaf->type = c_type; \
- leaf->level = c_level; \
- leaf->coherency_line_size = c->cache.linesz; \
- leaf->number_of_sets = c->cache.sets; \
- leaf->ways_of_associativity = c->cache.ways; \
- leaf->size = c->cache.linesz * c->cache.sets * \
- c->cache.ways; \
- leaf++; \
- } while (0)
- int init_cache_level(unsigned int cpu)
- {
- struct cpuinfo_mips *c = ¤t_cpu_data;
- struct cpu_cacheinfo *this_cpu_ci = get_cpu_cacheinfo(cpu);
- int levels = 0, leaves = 0;
- /*
- * If Dcache is not set, we assume the cache structures
- * are not properly initialized.
- */
- if (c->dcache.waysize)
- levels += 1;
- else
- return -ENOENT;
- leaves += (c->icache.waysize) ? 2 : 1;
- if (c->vcache.waysize) {
- levels++;
- leaves++;
- }
- if (c->scache.waysize) {
- levels++;
- leaves++;
- }
- if (c->tcache.waysize) {
- levels++;
- leaves++;
- }
- this_cpu_ci->num_levels = levels;
- this_cpu_ci->num_leaves = leaves;
- return 0;
- }
- static void fill_cpumask_siblings(int cpu, cpumask_t *cpu_map)
- {
- int cpu1;
- for_each_possible_cpu(cpu1)
- if (cpus_are_siblings(cpu, cpu1))
- cpumask_set_cpu(cpu1, cpu_map);
- }
- static void fill_cpumask_cluster(int cpu, cpumask_t *cpu_map)
- {
- int cpu1;
- int cluster = cpu_cluster(&cpu_data[cpu]);
- for_each_possible_cpu(cpu1)
- if (cpu_cluster(&cpu_data[cpu1]) == cluster)
- cpumask_set_cpu(cpu1, cpu_map);
- }
- int populate_cache_leaves(unsigned int cpu)
- {
- struct cpuinfo_mips *c = ¤t_cpu_data;
- struct cpu_cacheinfo *this_cpu_ci = get_cpu_cacheinfo(cpu);
- struct cacheinfo *this_leaf = this_cpu_ci->info_list;
- int level = 1;
- if (c->icache.waysize) {
- /* I/D caches are per core */
- fill_cpumask_siblings(cpu, &this_leaf->shared_cpu_map);
- populate_cache(dcache, this_leaf, level, CACHE_TYPE_DATA);
- fill_cpumask_siblings(cpu, &this_leaf->shared_cpu_map);
- populate_cache(icache, this_leaf, level, CACHE_TYPE_INST);
- level++;
- } else {
- populate_cache(dcache, this_leaf, level, CACHE_TYPE_UNIFIED);
- level++;
- }
- if (c->vcache.waysize) {
- /* Vcache is per core as well */
- fill_cpumask_siblings(cpu, &this_leaf->shared_cpu_map);
- populate_cache(vcache, this_leaf, level, CACHE_TYPE_UNIFIED);
- level++;
- }
- if (c->scache.waysize) {
- /* Scache is per cluster */
- fill_cpumask_cluster(cpu, &this_leaf->shared_cpu_map);
- populate_cache(scache, this_leaf, level, CACHE_TYPE_UNIFIED);
- level++;
- }
- if (c->tcache.waysize)
- populate_cache(tcache, this_leaf, level, CACHE_TYPE_UNIFIED);
- this_cpu_ci->cpu_map_populated = true;
- return 0;
- }
|