1 /* 2 * Extract CPU cache information and expose them via sysfs. 3 * 4 * Copyright IBM Corp. 2012 5 * Author(s): Heiko Carstens <heiko.carstens@de.ibm.com> 6 */ 7 8 #include <linux/seq_file.h> 9 #include <linux/cpu.h> 10 #include <linux/cacheinfo.h> 11 #include <asm/facility.h> 12 13 enum { 14 CACHE_SCOPE_NOTEXISTS, 15 CACHE_SCOPE_PRIVATE, 16 CACHE_SCOPE_SHARED, 17 CACHE_SCOPE_RESERVED, 18 }; 19 20 enum { 21 CTYPE_SEPARATE, 22 CTYPE_DATA, 23 CTYPE_INSTRUCTION, 24 CTYPE_UNIFIED, 25 }; 26 27 enum { 28 EXTRACT_TOPOLOGY, 29 EXTRACT_LINE_SIZE, 30 EXTRACT_SIZE, 31 EXTRACT_ASSOCIATIVITY, 32 }; 33 34 enum { 35 CACHE_TI_UNIFIED = 0, 36 CACHE_TI_DATA = 0, 37 CACHE_TI_INSTRUCTION, 38 }; 39 40 struct cache_info { 41 unsigned char : 4; 42 unsigned char scope : 2; 43 unsigned char type : 2; 44 }; 45 46 #define CACHE_MAX_LEVEL 8 47 union cache_topology { 48 struct cache_info ci[CACHE_MAX_LEVEL]; 49 unsigned long long raw; 50 }; 51 52 static const char * const cache_type_string[] = { 53 "", 54 "Instruction", 55 "Data", 56 "", 57 "Unified", 58 }; 59 60 static const enum cache_type cache_type_map[] = { 61 [CTYPE_SEPARATE] = CACHE_TYPE_SEPARATE, 62 [CTYPE_DATA] = CACHE_TYPE_DATA, 63 [CTYPE_INSTRUCTION] = CACHE_TYPE_INST, 64 [CTYPE_UNIFIED] = CACHE_TYPE_UNIFIED, 65 }; 66 67 void show_cacheinfo(struct seq_file *m) 68 { 69 struct cpu_cacheinfo *this_cpu_ci; 70 struct cacheinfo *cache; 71 int idx; 72 73 if (!test_facility(34)) 74 return; 75 this_cpu_ci = get_cpu_cacheinfo(cpumask_any(cpu_online_mask)); 76 for (idx = 0; idx < this_cpu_ci->num_leaves; idx++) { 77 cache = this_cpu_ci->info_list + idx; 78 seq_printf(m, "cache%-11d: ", idx); 79 seq_printf(m, "level=%d ", cache->level); 80 seq_printf(m, "type=%s ", cache_type_string[cache->type]); 81 seq_printf(m, "scope=%s ", 82 cache->disable_sysfs ? "Shared" : "Private"); 83 seq_printf(m, "size=%dK ", cache->size >> 10); 84 seq_printf(m, "line_size=%u ", cache->coherency_line_size); 85 seq_printf(m, "associativity=%d", cache->ways_of_associativity); 86 seq_puts(m, "\n"); 87 } 88 } 89 90 static inline enum cache_type get_cache_type(struct cache_info *ci, int level) 91 { 92 if (level >= CACHE_MAX_LEVEL) 93 return CACHE_TYPE_NOCACHE; 94 ci += level; 95 if (ci->scope != CACHE_SCOPE_SHARED && ci->scope != CACHE_SCOPE_PRIVATE) 96 return CACHE_TYPE_NOCACHE; 97 return cache_type_map[ci->type]; 98 } 99 100 static inline unsigned long ecag(int ai, int li, int ti) 101 { 102 unsigned long cmd, val; 103 104 cmd = ai << 4 | li << 1 | ti; 105 asm volatile(".insn rsy,0xeb000000004c,%0,0,0(%1)" /* ecag */ 106 : "=d" (val) : "a" (cmd)); 107 return val; 108 } 109 110 static void ci_leaf_init(struct cacheinfo *this_leaf, int private, 111 enum cache_type type, unsigned int level, int cpu) 112 { 113 int ti, num_sets; 114 115 if (type == CACHE_TYPE_INST) 116 ti = CACHE_TI_INSTRUCTION; 117 else 118 ti = CACHE_TI_UNIFIED; 119 this_leaf->level = level + 1; 120 this_leaf->type = type; 121 this_leaf->coherency_line_size = ecag(EXTRACT_LINE_SIZE, level, ti); 122 this_leaf->ways_of_associativity = ecag(EXTRACT_ASSOCIATIVITY, level, ti); 123 this_leaf->size = ecag(EXTRACT_SIZE, level, ti); 124 num_sets = this_leaf->size / this_leaf->coherency_line_size; 125 num_sets /= this_leaf->ways_of_associativity; 126 this_leaf->number_of_sets = num_sets; 127 cpumask_set_cpu(cpu, &this_leaf->shared_cpu_map); 128 if (!private) 129 this_leaf->disable_sysfs = true; 130 } 131 132 int init_cache_level(unsigned int cpu) 133 { 134 struct cpu_cacheinfo *this_cpu_ci = get_cpu_cacheinfo(cpu); 135 unsigned int level = 0, leaves = 0; 136 union cache_topology ct; 137 enum cache_type ctype; 138 139 if (!test_facility(34)) 140 return -EOPNOTSUPP; 141 if (!this_cpu_ci) 142 return -EINVAL; 143 ct.raw = ecag(EXTRACT_TOPOLOGY, 0, 0); 144 do { 145 ctype = get_cache_type(&ct.ci[0], level); 146 if (ctype == CACHE_TYPE_NOCACHE) 147 break; 148 /* Separate instruction and data caches */ 149 leaves += (ctype == CACHE_TYPE_SEPARATE) ? 2 : 1; 150 } while (++level < CACHE_MAX_LEVEL); 151 this_cpu_ci->num_levels = level; 152 this_cpu_ci->num_leaves = leaves; 153 return 0; 154 } 155 156 int populate_cache_leaves(unsigned int cpu) 157 { 158 struct cpu_cacheinfo *this_cpu_ci = get_cpu_cacheinfo(cpu); 159 struct cacheinfo *this_leaf = this_cpu_ci->info_list; 160 unsigned int level, idx, pvt; 161 union cache_topology ct; 162 enum cache_type ctype; 163 164 if (!test_facility(34)) 165 return -EOPNOTSUPP; 166 ct.raw = ecag(EXTRACT_TOPOLOGY, 0, 0); 167 for (idx = 0, level = 0; level < this_cpu_ci->num_levels && 168 idx < this_cpu_ci->num_leaves; idx++, level++) { 169 if (!this_leaf) 170 return -EINVAL; 171 pvt = (ct.ci[level].scope == CACHE_SCOPE_PRIVATE) ? 1 : 0; 172 ctype = get_cache_type(&ct.ci[0], level); 173 if (ctype == CACHE_TYPE_SEPARATE) { 174 ci_leaf_init(this_leaf++, pvt, CACHE_TYPE_DATA, level, cpu); 175 ci_leaf_init(this_leaf++, pvt, CACHE_TYPE_INST, level, cpu); 176 } else { 177 ci_leaf_init(this_leaf++, pvt, ctype, level, cpu); 178 } 179 } 180 return 0; 181 } 182