Loading...
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 236 237 238 239 240 241 242 243 244 245 246 247 248 249 250 251 252 253 | // SPDX-License-Identifier: GPL-2.0-or-later /* * Copyright 2016-17 IBM Corp. */ #define pr_fmt(fmt) "vas: " fmt #include <linux/module.h> #include <linux/kernel.h> #include <linux/export.h> #include <linux/types.h> #include <linux/slab.h> #include <linux/platform_device.h> #include <linux/of_platform.h> #include <linux/of_address.h> #include <linux/of.h> #include <linux/irqdomain.h> #include <linux/interrupt.h> #include <asm/prom.h> #include <asm/xive.h> #include "vas.h" DEFINE_MUTEX(vas_mutex); static LIST_HEAD(vas_instances); static DEFINE_PER_CPU(int, cpu_vas_id); static int vas_irq_fault_window_setup(struct vas_instance *vinst) { int rc = 0; rc = request_threaded_irq(vinst->virq, vas_fault_handler, vas_fault_thread_fn, 0, vinst->name, vinst); if (rc) { pr_err("VAS[%d]: Request IRQ(%d) failed with %d\n", vinst->vas_id, vinst->virq, rc); goto out; } rc = vas_setup_fault_window(vinst); if (rc) free_irq(vinst->virq, vinst); out: return rc; } static int init_vas_instance(struct platform_device *pdev) { struct device_node *dn = pdev->dev.of_node; struct vas_instance *vinst; struct xive_irq_data *xd; uint32_t chipid, hwirq; struct resource *res; int rc, cpu, vasid; rc = of_property_read_u32(dn, "ibm,vas-id", &vasid); if (rc) { pr_err("No ibm,vas-id property for %s?\n", pdev->name); return -ENODEV; } rc = of_property_read_u32(dn, "ibm,chip-id", &chipid); if (rc) { pr_err("No ibm,chip-id property for %s?\n", pdev->name); return -ENODEV; } if (pdev->num_resources != 4) { pr_err("Unexpected DT configuration for [%s, %d]\n", pdev->name, vasid); return -ENODEV; } vinst = kzalloc(sizeof(*vinst), GFP_KERNEL); if (!vinst) return -ENOMEM; vinst->name = kasprintf(GFP_KERNEL, "vas-%d", vasid); if (!vinst->name) { kfree(vinst); return -ENOMEM; } INIT_LIST_HEAD(&vinst->node); ida_init(&vinst->ida); mutex_init(&vinst->mutex); vinst->vas_id = vasid; vinst->pdev = pdev; res = &pdev->resource[0]; vinst->hvwc_bar_start = res->start; res = &pdev->resource[1]; vinst->uwc_bar_start = res->start; res = &pdev->resource[2]; vinst->paste_base_addr = res->start; res = &pdev->resource[3]; if (res->end > 62) { pr_err("Bad 'paste_win_id_shift' in DT, %llx\n", res->end); goto free_vinst; } vinst->paste_win_id_shift = 63 - res->end; hwirq = xive_native_alloc_irq_on_chip(chipid); if (!hwirq) { pr_err("Inst%d: Unable to allocate global irq for chip %d\n", vinst->vas_id, chipid); return -ENOENT; } vinst->virq = irq_create_mapping(NULL, hwirq); if (!vinst->virq) { pr_err("Inst%d: Unable to map global irq %d\n", vinst->vas_id, hwirq); return -EINVAL; } xd = irq_get_handler_data(vinst->virq); if (!xd) { pr_err("Inst%d: Invalid virq %d\n", vinst->vas_id, vinst->virq); return -EINVAL; } vinst->irq_port = xd->trig_page; pr_devel("Initialized instance [%s, %d] paste_base 0x%llx paste_win_id_shift 0x%llx IRQ %d Port 0x%llx\n", pdev->name, vasid, vinst->paste_base_addr, vinst->paste_win_id_shift, vinst->virq, vinst->irq_port); for_each_possible_cpu(cpu) { if (cpu_to_chip_id(cpu) == of_get_ibm_chip_id(dn)) per_cpu(cpu_vas_id, cpu) = vasid; } mutex_lock(&vas_mutex); list_add(&vinst->node, &vas_instances); mutex_unlock(&vas_mutex); spin_lock_init(&vinst->fault_lock); /* * IRQ and fault handling setup is needed only for user space * send windows. */ if (vinst->virq) { rc = vas_irq_fault_window_setup(vinst); /* * Fault window is used only for user space send windows. * So if vinst->virq is NULL, tx_win_open returns -ENODEV * for user space. */ if (rc) vinst->virq = 0; } vas_instance_init_dbgdir(vinst); dev_set_drvdata(&pdev->dev, vinst); return 0; free_vinst: kfree(vinst->name); kfree(vinst); return -ENODEV; } /* * Although this is read/used multiple times, it is written to only * during initialization. */ struct vas_instance *find_vas_instance(int vasid) { struct list_head *ent; struct vas_instance *vinst; mutex_lock(&vas_mutex); if (vasid == -1) vasid = per_cpu(cpu_vas_id, smp_processor_id()); list_for_each(ent, &vas_instances) { vinst = list_entry(ent, struct vas_instance, node); if (vinst->vas_id == vasid) { mutex_unlock(&vas_mutex); return vinst; } } mutex_unlock(&vas_mutex); pr_devel("Instance %d not found\n", vasid); return NULL; } int chip_to_vas_id(int chipid) { int cpu; for_each_possible_cpu(cpu) { if (cpu_to_chip_id(cpu) == chipid) return per_cpu(cpu_vas_id, cpu); } return -1; } EXPORT_SYMBOL(chip_to_vas_id); static int vas_probe(struct platform_device *pdev) { return init_vas_instance(pdev); } static const struct of_device_id powernv_vas_match[] = { { .compatible = "ibm,vas",}, {}, }; static struct platform_driver vas_driver = { .driver = { .name = "vas", .of_match_table = powernv_vas_match, }, .probe = vas_probe, }; static int __init vas_init(void) { int found = 0; struct device_node *dn; platform_driver_register(&vas_driver); for_each_compatible_node(dn, NULL, "ibm,vas") { of_platform_device_create(dn, NULL, NULL); found++; } if (!found) { platform_driver_unregister(&vas_driver); return -ENODEV; } pr_devel("Found %d instances\n", found); return 0; } device_initcall(vas_init); |