| 1 | // SPDX-License-Identifier: GPL-2.0-only |
| 2 | /* Copyright(c) 2022 Intel Corporation */ |
| 3 | #include <linux/module.h> |
| 4 | #include <linux/slab.h> |
| 5 | #include "adf_accel_devices.h" |
| 6 | #include "adf_common_drv.h" |
| 7 | #include "adf_transport.h" |
| 8 | #include "adf_transport_access_macros.h" |
| 9 | #include "adf_cfg.h" |
| 10 | #include "adf_cfg_strings.h" |
| 11 | #include "qat_compression.h" |
| 12 | #include "icp_qat_fw.h" |
| 13 | |
| 14 | #define SEC ADF_KERNEL_SEC |
| 15 | |
| 16 | static struct service_hndl qat_compression; |
| 17 | |
| 18 | void qat_compression_put_instance(struct qat_compression_instance *inst) |
| 19 | { |
| 20 | atomic_dec(v: &inst->refctr); |
| 21 | adf_dev_put(accel_dev: inst->accel_dev); |
| 22 | } |
| 23 | |
| 24 | static int qat_compression_free_instances(struct adf_accel_dev *accel_dev) |
| 25 | { |
| 26 | struct qat_compression_instance *inst; |
| 27 | struct list_head *list_ptr, *tmp; |
| 28 | int i; |
| 29 | |
| 30 | list_for_each_safe(list_ptr, tmp, &accel_dev->compression_list) { |
| 31 | inst = list_entry(list_ptr, |
| 32 | struct qat_compression_instance, list); |
| 33 | |
| 34 | for (i = 0; i < atomic_read(v: &inst->refctr); i++) |
| 35 | qat_compression_put_instance(inst); |
| 36 | |
| 37 | if (inst->dc_tx) |
| 38 | adf_remove_ring(ring: inst->dc_tx); |
| 39 | |
| 40 | if (inst->dc_rx) |
| 41 | adf_remove_ring(ring: inst->dc_rx); |
| 42 | |
| 43 | list_del(entry: list_ptr); |
| 44 | kfree(objp: inst); |
| 45 | } |
| 46 | return 0; |
| 47 | } |
| 48 | |
| 49 | struct qat_compression_instance *qat_compression_get_instance_node(int node) |
| 50 | { |
| 51 | struct qat_compression_instance *inst = NULL; |
| 52 | struct adf_accel_dev *accel_dev = NULL; |
| 53 | unsigned long best = ~0; |
| 54 | struct list_head *itr; |
| 55 | |
| 56 | list_for_each(itr, adf_devmgr_get_head()) { |
| 57 | struct adf_accel_dev *tmp_dev; |
| 58 | unsigned long ctr; |
| 59 | int tmp_dev_node; |
| 60 | |
| 61 | tmp_dev = list_entry(itr, struct adf_accel_dev, list); |
| 62 | tmp_dev_node = dev_to_node(dev: &GET_DEV(tmp_dev)); |
| 63 | |
| 64 | if ((node == tmp_dev_node || tmp_dev_node < 0) && |
| 65 | adf_dev_started(accel_dev: tmp_dev) && !list_empty(head: &tmp_dev->compression_list)) { |
| 66 | ctr = atomic_read(v: &tmp_dev->ref_count); |
| 67 | if (best > ctr) { |
| 68 | accel_dev = tmp_dev; |
| 69 | best = ctr; |
| 70 | } |
| 71 | } |
| 72 | } |
| 73 | |
| 74 | if (!accel_dev) { |
| 75 | pr_debug_ratelimited("QAT: Could not find a device on node %d\n" , node); |
| 76 | /* Get any started device */ |
| 77 | list_for_each(itr, adf_devmgr_get_head()) { |
| 78 | struct adf_accel_dev *tmp_dev; |
| 79 | |
| 80 | tmp_dev = list_entry(itr, struct adf_accel_dev, list); |
| 81 | if (adf_dev_started(accel_dev: tmp_dev) && |
| 82 | !list_empty(head: &tmp_dev->compression_list)) { |
| 83 | accel_dev = tmp_dev; |
| 84 | break; |
| 85 | } |
| 86 | } |
| 87 | } |
| 88 | |
| 89 | if (!accel_dev) |
| 90 | return NULL; |
| 91 | |
| 92 | best = ~0; |
| 93 | list_for_each(itr, &accel_dev->compression_list) { |
| 94 | struct qat_compression_instance *tmp_inst; |
| 95 | unsigned long ctr; |
| 96 | |
| 97 | tmp_inst = list_entry(itr, struct qat_compression_instance, list); |
| 98 | ctr = atomic_read(v: &tmp_inst->refctr); |
| 99 | if (best > ctr) { |
| 100 | inst = tmp_inst; |
| 101 | best = ctr; |
| 102 | } |
| 103 | } |
| 104 | if (inst) { |
| 105 | if (adf_dev_get(accel_dev)) { |
| 106 | dev_err(&GET_DEV(accel_dev), "Could not increment dev refctr\n" ); |
| 107 | return NULL; |
| 108 | } |
| 109 | atomic_inc(v: &inst->refctr); |
| 110 | } |
| 111 | return inst; |
| 112 | } |
| 113 | |
| 114 | static int qat_compression_create_instances(struct adf_accel_dev *accel_dev) |
| 115 | { |
| 116 | struct qat_compression_instance *inst; |
| 117 | char key[ADF_CFG_MAX_KEY_LEN_IN_BYTES]; |
| 118 | char val[ADF_CFG_MAX_VAL_LEN_IN_BYTES]; |
| 119 | unsigned long num_inst, num_msg_dc; |
| 120 | unsigned long bank; |
| 121 | int msg_size; |
| 122 | int ret; |
| 123 | int i; |
| 124 | |
| 125 | INIT_LIST_HEAD(list: &accel_dev->compression_list); |
| 126 | strscpy(key, ADF_NUM_DC, sizeof(key)); |
| 127 | ret = adf_cfg_get_param_value(accel_dev, SEC, name: key, value: val); |
| 128 | if (ret) |
| 129 | return ret; |
| 130 | |
| 131 | ret = kstrtoul(s: val, base: 10, res: &num_inst); |
| 132 | if (ret) |
| 133 | return ret; |
| 134 | |
| 135 | for (i = 0; i < num_inst; i++) { |
| 136 | inst = kzalloc_node(sizeof(*inst), GFP_KERNEL, |
| 137 | dev_to_node(&GET_DEV(accel_dev))); |
| 138 | if (!inst) { |
| 139 | ret = -ENOMEM; |
| 140 | goto err; |
| 141 | } |
| 142 | |
| 143 | list_add_tail(new: &inst->list, head: &accel_dev->compression_list); |
| 144 | inst->id = i; |
| 145 | atomic_set(v: &inst->refctr, i: 0); |
| 146 | inst->accel_dev = accel_dev; |
| 147 | |
| 148 | snprintf(buf: key, size: sizeof(key), ADF_DC "%d" ADF_RING_DC_BANK_NUM, i); |
| 149 | ret = adf_cfg_get_param_value(accel_dev, SEC, name: key, value: val); |
| 150 | if (ret) |
| 151 | return ret; |
| 152 | |
| 153 | ret = kstrtoul(s: val, base: 10, res: &bank); |
| 154 | if (ret) |
| 155 | return ret; |
| 156 | |
| 157 | snprintf(buf: key, size: sizeof(key), ADF_DC "%d" ADF_RING_DC_SIZE, i); |
| 158 | ret = adf_cfg_get_param_value(accel_dev, SEC, name: key, value: val); |
| 159 | if (ret) |
| 160 | return ret; |
| 161 | |
| 162 | ret = kstrtoul(s: val, base: 10, res: &num_msg_dc); |
| 163 | if (ret) |
| 164 | return ret; |
| 165 | |
| 166 | msg_size = ICP_QAT_FW_REQ_DEFAULT_SZ; |
| 167 | snprintf(buf: key, size: sizeof(key), ADF_DC "%d" ADF_RING_DC_TX, i); |
| 168 | ret = adf_create_ring(accel_dev, SEC, bank_num: bank, num_mgs: num_msg_dc, |
| 169 | msg_size, ring_name: key, NULL, poll_mode: 0, ring_ptr: &inst->dc_tx); |
| 170 | if (ret) |
| 171 | return ret; |
| 172 | |
| 173 | msg_size = ICP_QAT_FW_RESP_DEFAULT_SZ; |
| 174 | snprintf(buf: key, size: sizeof(key), ADF_DC "%d" ADF_RING_DC_RX, i); |
| 175 | ret = adf_create_ring(accel_dev, SEC, bank_num: bank, num_mgs: num_msg_dc, |
| 176 | msg_size, ring_name: key, callback: qat_comp_alg_callback, poll_mode: 0, |
| 177 | ring_ptr: &inst->dc_rx); |
| 178 | if (ret) |
| 179 | return ret; |
| 180 | |
| 181 | inst->dc_data = accel_dev->dc_data; |
| 182 | INIT_LIST_HEAD(list: &inst->backlog.list); |
| 183 | spin_lock_init(&inst->backlog.lock); |
| 184 | } |
| 185 | return 0; |
| 186 | err: |
| 187 | qat_compression_free_instances(accel_dev); |
| 188 | return ret; |
| 189 | } |
| 190 | |
| 191 | static int qat_compression_alloc_dc_data(struct adf_accel_dev *accel_dev) |
| 192 | { |
| 193 | struct device *dev = &GET_DEV(accel_dev); |
| 194 | dma_addr_t obuff_p = DMA_MAPPING_ERROR; |
| 195 | size_t ovf_buff_sz = QAT_COMP_MAX_SKID; |
| 196 | struct adf_dc_data *dc_data = NULL; |
| 197 | u8 *obuff = NULL; |
| 198 | |
| 199 | dc_data = kzalloc_node(sizeof(*dc_data), GFP_KERNEL, dev_to_node(dev)); |
| 200 | if (!dc_data) |
| 201 | goto err; |
| 202 | |
| 203 | obuff = kzalloc_node(ovf_buff_sz, GFP_KERNEL, dev_to_node(dev)); |
| 204 | if (!obuff) |
| 205 | goto err; |
| 206 | |
| 207 | obuff_p = dma_map_single(dev, obuff, ovf_buff_sz, DMA_BIDIRECTIONAL); |
| 208 | if (unlikely(dma_mapping_error(dev, obuff_p))) |
| 209 | goto err; |
| 210 | |
| 211 | dc_data->ovf_buff = obuff; |
| 212 | dc_data->ovf_buff_p = obuff_p; |
| 213 | dc_data->ovf_buff_sz = ovf_buff_sz; |
| 214 | |
| 215 | accel_dev->dc_data = dc_data; |
| 216 | |
| 217 | return 0; |
| 218 | |
| 219 | err: |
| 220 | accel_dev->dc_data = NULL; |
| 221 | kfree(objp: obuff); |
| 222 | devm_kfree(dev, p: dc_data); |
| 223 | return -ENOMEM; |
| 224 | } |
| 225 | |
| 226 | static void qat_free_dc_data(struct adf_accel_dev *accel_dev) |
| 227 | { |
| 228 | struct adf_dc_data *dc_data = accel_dev->dc_data; |
| 229 | struct device *dev = &GET_DEV(accel_dev); |
| 230 | |
| 231 | if (!dc_data) |
| 232 | return; |
| 233 | |
| 234 | dma_unmap_single(dev, dc_data->ovf_buff_p, dc_data->ovf_buff_sz, |
| 235 | DMA_BIDIRECTIONAL); |
| 236 | kfree_sensitive(objp: dc_data->ovf_buff); |
| 237 | kfree(objp: dc_data); |
| 238 | accel_dev->dc_data = NULL; |
| 239 | } |
| 240 | |
| 241 | static int qat_compression_init(struct adf_accel_dev *accel_dev) |
| 242 | { |
| 243 | int ret; |
| 244 | |
| 245 | ret = qat_compression_alloc_dc_data(accel_dev); |
| 246 | if (ret) |
| 247 | return ret; |
| 248 | |
| 249 | ret = qat_compression_create_instances(accel_dev); |
| 250 | if (ret) |
| 251 | qat_free_dc_data(accel_dev); |
| 252 | |
| 253 | return ret; |
| 254 | } |
| 255 | |
| 256 | static int qat_compression_shutdown(struct adf_accel_dev *accel_dev) |
| 257 | { |
| 258 | qat_free_dc_data(accel_dev); |
| 259 | return qat_compression_free_instances(accel_dev); |
| 260 | } |
| 261 | |
| 262 | static int qat_compression_event_handler(struct adf_accel_dev *accel_dev, |
| 263 | enum adf_event event) |
| 264 | { |
| 265 | int ret; |
| 266 | |
| 267 | switch (event) { |
| 268 | case ADF_EVENT_INIT: |
| 269 | ret = qat_compression_init(accel_dev); |
| 270 | break; |
| 271 | case ADF_EVENT_SHUTDOWN: |
| 272 | ret = qat_compression_shutdown(accel_dev); |
| 273 | break; |
| 274 | case ADF_EVENT_RESTARTING: |
| 275 | case ADF_EVENT_RESTARTED: |
| 276 | case ADF_EVENT_START: |
| 277 | case ADF_EVENT_STOP: |
| 278 | default: |
| 279 | ret = 0; |
| 280 | } |
| 281 | return ret; |
| 282 | } |
| 283 | |
| 284 | int qat_compression_register(void) |
| 285 | { |
| 286 | memset(&qat_compression, 0, sizeof(qat_compression)); |
| 287 | qat_compression.event_hld = qat_compression_event_handler; |
| 288 | qat_compression.name = "qat_compression" ; |
| 289 | return adf_service_register(service: &qat_compression); |
| 290 | } |
| 291 | |
| 292 | int qat_compression_unregister(void) |
| 293 | { |
| 294 | return adf_service_unregister(service: &qat_compression); |
| 295 | } |
| 296 | |