Loading...
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 236 237 238 239 240 241 242 243 244 245 246 247 248 249 250 251 252 253 254 255 256 257 258 259 260 261 262 263 264 265 266 267 268 269 270 271 272 273 274 275 276 277 278 279 280 281 282 283 284 285 286 287 288 289 290 291 292 293 294 295 296 297 298 299 | // SPDX-License-Identifier: MIT /* * Copyright © 2017 Intel Corporation */ #include "i915_selftest.h" #include "gem/i915_gem_pm.h" /* max doorbell number + negative test for each client type */ #define ATTEMPTS (GUC_NUM_DOORBELLS + GUC_CLIENT_PRIORITY_NUM) static struct intel_guc_client *clients[ATTEMPTS]; static bool available_dbs(struct intel_guc *guc, u32 priority) { unsigned long offset; unsigned long end; u16 id; /* first half is used for normal priority, second half for high */ offset = 0; end = GUC_NUM_DOORBELLS / 2; if (priority <= GUC_CLIENT_PRIORITY_HIGH) { offset = end; end += offset; } id = find_next_zero_bit(guc->doorbell_bitmap, end, offset); if (id < end) return true; return false; } static int check_all_doorbells(struct intel_guc *guc) { u16 db_id; pr_info_once("Max number of doorbells: %d", GUC_NUM_DOORBELLS); for (db_id = 0; db_id < GUC_NUM_DOORBELLS; ++db_id) { if (!doorbell_ok(guc, db_id)) { pr_err("doorbell %d, not ok\n", db_id); return -EIO; } } return 0; } static int ring_doorbell_nop(struct intel_guc_client *client) { struct guc_process_desc *desc = __get_process_desc(client); int err; client->use_nop_wqi = true; spin_lock_irq(&client->wq_lock); guc_wq_item_append(client, 0, 0, 0, 0); guc_ring_doorbell(client); spin_unlock_irq(&client->wq_lock); client->use_nop_wqi = false; /* if there are no issues GuC will update the WQ head and keep the * WQ in active status */ err = wait_for(READ_ONCE(desc->head) == READ_ONCE(desc->tail), 10); if (err) { pr_err("doorbell %u ring failed!\n", client->doorbell_id); return -EIO; } if (desc->wq_status != WQ_STATUS_ACTIVE) { pr_err("doorbell %u ring put WQ in bad state (%u)!\n", client->doorbell_id, desc->wq_status); return -EIO; } return 0; } /* * Basic client sanity check, handy to validate create_clients. */ static int validate_client(struct intel_guc_client *client, int client_priority) { if (client->priority != client_priority || client->doorbell_id == GUC_DOORBELL_INVALID) return -EINVAL; else return 0; } static bool client_doorbell_in_sync(struct intel_guc_client *client) { return !client || doorbell_ok(client->guc, client->doorbell_id); } /* * Check that we're able to synchronize guc_clients with their doorbells * * We're creating clients and reserving doorbells once, at module load. During * module lifetime, GuC, doorbell HW, and i915 state may go out of sync due to * GuC being reset. In other words - GuC clients are still around, but the * status of their doorbells may be incorrect. This is the reason behind * validating that the doorbells status expected by the driver matches what the * GuC/HW have. */ static int igt_guc_clients(void *arg) { struct intel_gt *gt = arg; struct intel_guc *guc = >->uc.guc; intel_wakeref_t wakeref; int err = 0; GEM_BUG_ON(!HAS_GT_UC(gt->i915)); wakeref = intel_runtime_pm_get(gt->uncore->rpm); err = check_all_doorbells(guc); if (err) goto unlock; /* * Get rid of clients created during driver load because the test will * recreate them. */ guc_clients_disable(guc); guc_clients_destroy(guc); if (guc->execbuf_client) { pr_err("guc_clients_destroy lied!\n"); err = -EINVAL; goto unlock; } err = guc_clients_create(guc); if (err) { pr_err("Failed to create clients\n"); goto unlock; } GEM_BUG_ON(!guc->execbuf_client); err = validate_client(guc->execbuf_client, GUC_CLIENT_PRIORITY_KMD_NORMAL); if (err) { pr_err("execbug client validation failed\n"); goto out; } /* the client should now have reserved a doorbell */ if (!has_doorbell(guc->execbuf_client)) { pr_err("guc_clients_create didn't reserve doorbells\n"); err = -EINVAL; goto out; } /* Now enable the clients */ guc_clients_enable(guc); /* each client should now have received a doorbell */ if (!client_doorbell_in_sync(guc->execbuf_client)) { pr_err("failed to initialize the doorbells\n"); err = -EINVAL; goto out; } /* * Basic test - an attempt to reallocate a valid doorbell to the * client it is currently assigned should not cause a failure. */ err = create_doorbell(guc->execbuf_client); out: /* * Leave clean state for other test, plus the driver always destroy the * clients during unload. */ guc_clients_disable(guc); guc_clients_destroy(guc); guc_clients_create(guc); guc_clients_enable(guc); unlock: intel_runtime_pm_put(gt->uncore->rpm, wakeref); return err; } /* * Create as many clients as number of doorbells. Note that there's already * client(s)/doorbell(s) created during driver load, but this test creates * its own and do not interact with the existing ones. */ static int igt_guc_doorbells(void *arg) { struct intel_gt *gt = arg; struct intel_guc *guc = >->uc.guc; intel_wakeref_t wakeref; int i, err = 0; u16 db_id; GEM_BUG_ON(!HAS_GT_UC(gt->i915)); wakeref = intel_runtime_pm_get(gt->uncore->rpm); err = check_all_doorbells(guc); if (err) goto unlock; for (i = 0; i < ATTEMPTS; i++) { clients[i] = guc_client_alloc(guc, i % GUC_CLIENT_PRIORITY_NUM); if (!clients[i]) { pr_err("[%d] No guc client\n", i); err = -EINVAL; goto out; } if (IS_ERR(clients[i])) { if (PTR_ERR(clients[i]) != -ENOSPC) { pr_err("[%d] unexpected error\n", i); err = PTR_ERR(clients[i]); goto out; } if (available_dbs(guc, i % GUC_CLIENT_PRIORITY_NUM)) { pr_err("[%d] non-db related alloc fail\n", i); err = -EINVAL; goto out; } /* expected, ran out of dbs for this client type */ continue; } /* * The check below is only valid because we keep a doorbell * assigned during the whole life of the client. */ if (clients[i]->stage_id >= GUC_NUM_DOORBELLS) { pr_err("[%d] more clients than doorbells (%d >= %d)\n", i, clients[i]->stage_id, GUC_NUM_DOORBELLS); err = -EINVAL; goto out; } err = validate_client(clients[i], i % GUC_CLIENT_PRIORITY_NUM); if (err) { pr_err("[%d] client_alloc sanity check failed!\n", i); err = -EINVAL; goto out; } db_id = clients[i]->doorbell_id; err = __guc_client_enable(clients[i]); if (err) { pr_err("[%d] Failed to create a doorbell\n", i); goto out; } /* doorbell id shouldn't change, we are holding the mutex */ if (db_id != clients[i]->doorbell_id) { pr_err("[%d] doorbell id changed (%d != %d)\n", i, db_id, clients[i]->doorbell_id); err = -EINVAL; goto out; } err = check_all_doorbells(guc); if (err) goto out; err = ring_doorbell_nop(clients[i]); if (err) goto out; } out: for (i = 0; i < ATTEMPTS; i++) if (!IS_ERR_OR_NULL(clients[i])) { __guc_client_disable(clients[i]); guc_client_free(clients[i]); } unlock: intel_runtime_pm_put(gt->uncore->rpm, wakeref); return err; } int intel_guc_live_selftest(struct drm_i915_private *i915) { static const struct i915_subtest tests[] = { SUBTEST(igt_guc_clients), SUBTEST(igt_guc_doorbells), }; if (!USES_GUC_SUBMISSION(i915)) return 0; return intel_gt_live_subtests(tests, &i915->gt); } |