| 1 | // SPDX-License-Identifier: GPL-2.0 |
| 2 | /* |
| 3 | * AMD L3 cache_disable_{0,1} sysfs handling |
| 4 | * Documentation/ABI/testing/sysfs-devices-system-cpu |
| 5 | */ |
| 6 | |
| 7 | #include <linux/cacheinfo.h> |
| 8 | #include <linux/capability.h> |
| 9 | #include <linux/pci.h> |
| 10 | #include <linux/sysfs.h> |
| 11 | |
| 12 | #include <asm/amd/nb.h> |
| 13 | |
| 14 | #include "cpu.h" |
| 15 | |
| 16 | /* |
| 17 | * L3 cache descriptors |
| 18 | */ |
| 19 | static void amd_calc_l3_indices(struct amd_northbridge *nb) |
| 20 | { |
| 21 | struct amd_l3_cache *l3 = &nb->l3_cache; |
| 22 | unsigned int sc0, sc1, sc2, sc3; |
| 23 | u32 val = 0; |
| 24 | |
| 25 | pci_read_config_dword(dev: nb->misc, where: 0x1C4, val: &val); |
| 26 | |
| 27 | /* calculate subcache sizes */ |
| 28 | l3->subcaches[0] = sc0 = !(val & BIT(0)); |
| 29 | l3->subcaches[1] = sc1 = !(val & BIT(4)); |
| 30 | |
| 31 | if (boot_cpu_data.x86 == 0x15) { |
| 32 | l3->subcaches[0] = sc0 += !(val & BIT(1)); |
| 33 | l3->subcaches[1] = sc1 += !(val & BIT(5)); |
| 34 | } |
| 35 | |
| 36 | l3->subcaches[2] = sc2 = !(val & BIT(8)) + !(val & BIT(9)); |
| 37 | l3->subcaches[3] = sc3 = !(val & BIT(12)) + !(val & BIT(13)); |
| 38 | |
| 39 | l3->indices = (max(max3(sc0, sc1, sc2), sc3) << 10) - 1; |
| 40 | } |
| 41 | |
| 42 | /* |
| 43 | * check whether a slot used for disabling an L3 index is occupied. |
| 44 | * @l3: L3 cache descriptor |
| 45 | * @slot: slot number (0..1) |
| 46 | * |
| 47 | * @returns: the disabled index if used or negative value if slot free. |
| 48 | */ |
| 49 | static int amd_get_l3_disable_slot(struct amd_northbridge *nb, unsigned int slot) |
| 50 | { |
| 51 | unsigned int reg = 0; |
| 52 | |
| 53 | pci_read_config_dword(dev: nb->misc, where: 0x1BC + slot * 4, val: ®); |
| 54 | |
| 55 | /* check whether this slot is activated already */ |
| 56 | if (reg & (3UL << 30)) |
| 57 | return reg & 0xfff; |
| 58 | |
| 59 | return -1; |
| 60 | } |
| 61 | |
| 62 | static ssize_t show_cache_disable(struct cacheinfo *ci, char *buf, unsigned int slot) |
| 63 | { |
| 64 | int index; |
| 65 | struct amd_northbridge *nb = ci->priv; |
| 66 | |
| 67 | index = amd_get_l3_disable_slot(nb, slot); |
| 68 | if (index >= 0) |
| 69 | return sysfs_emit(buf, fmt: "%d\n" , index); |
| 70 | |
| 71 | return sysfs_emit(buf, fmt: "FREE\n" ); |
| 72 | } |
| 73 | |
| 74 | #define SHOW_CACHE_DISABLE(slot) \ |
| 75 | static ssize_t \ |
| 76 | cache_disable_##slot##_show(struct device *dev, \ |
| 77 | struct device_attribute *attr, char *buf) \ |
| 78 | { \ |
| 79 | struct cacheinfo *ci = dev_get_drvdata(dev); \ |
| 80 | return show_cache_disable(ci, buf, slot); \ |
| 81 | } |
| 82 | |
| 83 | SHOW_CACHE_DISABLE(0) |
| 84 | SHOW_CACHE_DISABLE(1) |
| 85 | |
| 86 | static void amd_l3_disable_index(struct amd_northbridge *nb, int cpu, |
| 87 | unsigned int slot, unsigned long idx) |
| 88 | { |
| 89 | int i; |
| 90 | |
| 91 | idx |= BIT(30); |
| 92 | |
| 93 | /* |
| 94 | * disable index in all 4 subcaches |
| 95 | */ |
| 96 | for (i = 0; i < 4; i++) { |
| 97 | u32 reg = idx | (i << 20); |
| 98 | |
| 99 | if (!nb->l3_cache.subcaches[i]) |
| 100 | continue; |
| 101 | |
| 102 | pci_write_config_dword(dev: nb->misc, where: 0x1BC + slot * 4, val: reg); |
| 103 | |
| 104 | /* |
| 105 | * We need to WBINVD on a core on the node containing the L3 |
| 106 | * cache which indices we disable therefore a simple wbinvd() |
| 107 | * is not sufficient. |
| 108 | */ |
| 109 | wbinvd_on_cpu(cpu); |
| 110 | |
| 111 | reg |= BIT(31); |
| 112 | pci_write_config_dword(dev: nb->misc, where: 0x1BC + slot * 4, val: reg); |
| 113 | } |
| 114 | } |
| 115 | |
| 116 | /* |
| 117 | * disable a L3 cache index by using a disable-slot |
| 118 | * |
| 119 | * @l3: L3 cache descriptor |
| 120 | * @cpu: A CPU on the node containing the L3 cache |
| 121 | * @slot: slot number (0..1) |
| 122 | * @index: index to disable |
| 123 | * |
| 124 | * @return: 0 on success, error status on failure |
| 125 | */ |
| 126 | static int amd_set_l3_disable_slot(struct amd_northbridge *nb, int cpu, |
| 127 | unsigned int slot, unsigned long index) |
| 128 | { |
| 129 | int ret = 0; |
| 130 | |
| 131 | /* check if @slot is already used or the index is already disabled */ |
| 132 | ret = amd_get_l3_disable_slot(nb, slot); |
| 133 | if (ret >= 0) |
| 134 | return -EEXIST; |
| 135 | |
| 136 | if (index > nb->l3_cache.indices) |
| 137 | return -EINVAL; |
| 138 | |
| 139 | /* check whether the other slot has disabled the same index already */ |
| 140 | if (index == amd_get_l3_disable_slot(nb, slot: !slot)) |
| 141 | return -EEXIST; |
| 142 | |
| 143 | amd_l3_disable_index(nb, cpu, slot, idx: index); |
| 144 | |
| 145 | return 0; |
| 146 | } |
| 147 | |
| 148 | static ssize_t store_cache_disable(struct cacheinfo *ci, const char *buf, |
| 149 | size_t count, unsigned int slot) |
| 150 | { |
| 151 | struct amd_northbridge *nb = ci->priv; |
| 152 | unsigned long val = 0; |
| 153 | int cpu, err = 0; |
| 154 | |
| 155 | if (!capable(CAP_SYS_ADMIN)) |
| 156 | return -EPERM; |
| 157 | |
| 158 | cpu = cpumask_first(srcp: &ci->shared_cpu_map); |
| 159 | |
| 160 | if (kstrtoul(s: buf, base: 10, res: &val) < 0) |
| 161 | return -EINVAL; |
| 162 | |
| 163 | err = amd_set_l3_disable_slot(nb, cpu, slot, index: val); |
| 164 | if (err) { |
| 165 | if (err == -EEXIST) |
| 166 | pr_warn("L3 slot %d in use/index already disabled!\n" , |
| 167 | slot); |
| 168 | return err; |
| 169 | } |
| 170 | return count; |
| 171 | } |
| 172 | |
| 173 | #define STORE_CACHE_DISABLE(slot) \ |
| 174 | static ssize_t \ |
| 175 | cache_disable_##slot##_store(struct device *dev, \ |
| 176 | struct device_attribute *attr, \ |
| 177 | const char *buf, size_t count) \ |
| 178 | { \ |
| 179 | struct cacheinfo *ci = dev_get_drvdata(dev); \ |
| 180 | return store_cache_disable(ci, buf, count, slot); \ |
| 181 | } |
| 182 | |
| 183 | STORE_CACHE_DISABLE(0) |
| 184 | STORE_CACHE_DISABLE(1) |
| 185 | |
| 186 | static ssize_t subcaches_show(struct device *dev, struct device_attribute *attr, |
| 187 | char *buf) |
| 188 | { |
| 189 | struct cacheinfo *ci = dev_get_drvdata(dev); |
| 190 | int cpu = cpumask_first(srcp: &ci->shared_cpu_map); |
| 191 | |
| 192 | return sysfs_emit(buf, fmt: "%x\n" , amd_get_subcaches(cpu)); |
| 193 | } |
| 194 | |
| 195 | static ssize_t subcaches_store(struct device *dev, |
| 196 | struct device_attribute *attr, |
| 197 | const char *buf, size_t count) |
| 198 | { |
| 199 | struct cacheinfo *ci = dev_get_drvdata(dev); |
| 200 | int cpu = cpumask_first(srcp: &ci->shared_cpu_map); |
| 201 | unsigned long val; |
| 202 | |
| 203 | if (!capable(CAP_SYS_ADMIN)) |
| 204 | return -EPERM; |
| 205 | |
| 206 | if (kstrtoul(s: buf, base: 16, res: &val) < 0) |
| 207 | return -EINVAL; |
| 208 | |
| 209 | if (amd_set_subcaches(cpu, val)) |
| 210 | return -EINVAL; |
| 211 | |
| 212 | return count; |
| 213 | } |
| 214 | |
| 215 | static DEVICE_ATTR_RW(cache_disable_0); |
| 216 | static DEVICE_ATTR_RW(cache_disable_1); |
| 217 | static DEVICE_ATTR_RW(subcaches); |
| 218 | |
| 219 | static umode_t cache_private_attrs_is_visible(struct kobject *kobj, |
| 220 | struct attribute *attr, int unused) |
| 221 | { |
| 222 | struct device *dev = kobj_to_dev(kobj); |
| 223 | struct cacheinfo *ci = dev_get_drvdata(dev); |
| 224 | umode_t mode = attr->mode; |
| 225 | |
| 226 | if (!ci->priv) |
| 227 | return 0; |
| 228 | |
| 229 | if ((attr == &dev_attr_subcaches.attr) && |
| 230 | amd_nb_has_feature(AMD_NB_L3_PARTITIONING)) |
| 231 | return mode; |
| 232 | |
| 233 | if ((attr == &dev_attr_cache_disable_0.attr || |
| 234 | attr == &dev_attr_cache_disable_1.attr) && |
| 235 | amd_nb_has_feature(AMD_NB_L3_INDEX_DISABLE)) |
| 236 | return mode; |
| 237 | |
| 238 | return 0; |
| 239 | } |
| 240 | |
| 241 | static struct attribute_group cache_private_group = { |
| 242 | .is_visible = cache_private_attrs_is_visible, |
| 243 | }; |
| 244 | |
| 245 | static void init_amd_l3_attrs(void) |
| 246 | { |
| 247 | static struct attribute **amd_l3_attrs; |
| 248 | int n = 1; |
| 249 | |
| 250 | if (amd_l3_attrs) /* already initialized */ |
| 251 | return; |
| 252 | |
| 253 | if (amd_nb_has_feature(AMD_NB_L3_INDEX_DISABLE)) |
| 254 | n += 2; |
| 255 | if (amd_nb_has_feature(AMD_NB_L3_PARTITIONING)) |
| 256 | n += 1; |
| 257 | |
| 258 | amd_l3_attrs = kcalloc(n, sizeof(*amd_l3_attrs), GFP_KERNEL); |
| 259 | if (!amd_l3_attrs) |
| 260 | return; |
| 261 | |
| 262 | n = 0; |
| 263 | if (amd_nb_has_feature(AMD_NB_L3_INDEX_DISABLE)) { |
| 264 | amd_l3_attrs[n++] = &dev_attr_cache_disable_0.attr; |
| 265 | amd_l3_attrs[n++] = &dev_attr_cache_disable_1.attr; |
| 266 | } |
| 267 | if (amd_nb_has_feature(AMD_NB_L3_PARTITIONING)) |
| 268 | amd_l3_attrs[n++] = &dev_attr_subcaches.attr; |
| 269 | |
| 270 | cache_private_group.attrs = amd_l3_attrs; |
| 271 | } |
| 272 | |
| 273 | const struct attribute_group *cache_get_priv_group(struct cacheinfo *ci) |
| 274 | { |
| 275 | struct amd_northbridge *nb = ci->priv; |
| 276 | |
| 277 | if (ci->level < 3 || !nb) |
| 278 | return NULL; |
| 279 | |
| 280 | if (nb && nb->l3_cache.indices) |
| 281 | init_amd_l3_attrs(); |
| 282 | |
| 283 | return &cache_private_group; |
| 284 | } |
| 285 | |
| 286 | struct amd_northbridge *amd_init_l3_cache(int index) |
| 287 | { |
| 288 | struct amd_northbridge *nb; |
| 289 | int node; |
| 290 | |
| 291 | /* only for L3, and not in virtualized environments */ |
| 292 | if (index < 3) |
| 293 | return NULL; |
| 294 | |
| 295 | node = topology_amd_node_id(smp_processor_id()); |
| 296 | nb = node_to_amd_nb(node); |
| 297 | if (nb && !nb->l3_cache.indices) |
| 298 | amd_calc_l3_indices(nb); |
| 299 | |
| 300 | return nb; |
| 301 | } |
| 302 | |