1 /* $NetBSD: kfd_iommu.c,v 1.2 2021/12/18 23:44:59 riastradh Exp $ */ 2 3 /* 4 * Copyright 2018 Advanced Micro Devices, Inc. 5 * 6 * Permission is hereby granted, free of charge, to any person obtaining a 7 * copy of this software and associated documentation files (the "Software"), 8 * to deal in the Software without restriction, including without limitation 9 * the rights to use, copy, modify, merge, publish, distribute, sublicense, 10 * and/or sell copies of the Software, and to permit persons to whom the 11 * Software is furnished to do so, subject to the following conditions: 12 * 13 * The above copyright notice and this permission notice shall be included in 14 * all copies or substantial portions of the Software. 15 * 16 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR 17 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, 18 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL 19 * THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, DAMAGES OR 20 * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, 21 * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR 22 * OTHER DEALINGS IN THE SOFTWARE. 23 */ 24 25 #include <sys/cdefs.h> 26 __KERNEL_RCSID(0, "$NetBSD: kfd_iommu.c,v 1.2 2021/12/18 23:44:59 riastradh Exp $"); 27 28 #include <linux/printk.h> 29 #include <linux/device.h> 30 #include <linux/slab.h> 31 #include <linux/pci.h> 32 #include <linux/amd-iommu.h> 33 #include "kfd_priv.h" 34 #include "kfd_dbgmgr.h" 35 #include "kfd_topology.h" 36 #include "kfd_iommu.h" 37 38 static const u32 required_iommu_flags = AMD_IOMMU_DEVICE_FLAG_ATS_SUP | 39 AMD_IOMMU_DEVICE_FLAG_PRI_SUP | 40 AMD_IOMMU_DEVICE_FLAG_PASID_SUP; 41 42 /** kfd_iommu_check_device - Check whether IOMMU is available for device 43 */ 44 int kfd_iommu_check_device(struct kfd_dev *kfd) 45 { 46 struct amd_iommu_device_info iommu_info; 47 int err; 48 49 if (!kfd->device_info->needs_iommu_device) 50 return -ENODEV; 51 52 iommu_info.flags = 0; 53 err = amd_iommu_device_info(kfd->pdev, &iommu_info); 54 if (err) 55 return err; 56 57 if ((iommu_info.flags & required_iommu_flags) != required_iommu_flags) 58 return -ENODEV; 59 60 return 0; 61 } 62 63 /** kfd_iommu_device_init - Initialize IOMMU for device 64 */ 65 int kfd_iommu_device_init(struct kfd_dev *kfd) 66 { 67 struct amd_iommu_device_info iommu_info; 68 unsigned int pasid_limit; 69 int err; 70 71 if (!kfd->device_info->needs_iommu_device) 72 return 0; 73 74 iommu_info.flags = 0; 75 err = amd_iommu_device_info(kfd->pdev, &iommu_info); 76 if (err < 0) { 77 dev_err(kfd_device, 78 "error getting iommu info. is the iommu enabled?\n"); 79 return -ENODEV; 80 } 81 82 if ((iommu_info.flags & required_iommu_flags) != required_iommu_flags) { 83 dev_err(kfd_device, 84 "error required iommu flags ats %i, pri %i, pasid %i\n", 85 (iommu_info.flags & AMD_IOMMU_DEVICE_FLAG_ATS_SUP) != 0, 86 (iommu_info.flags & AMD_IOMMU_DEVICE_FLAG_PRI_SUP) != 0, 87 (iommu_info.flags & AMD_IOMMU_DEVICE_FLAG_PASID_SUP) 88 != 0); 89 return -ENODEV; 90 } 91 92 pasid_limit = min_t(unsigned int, 93 (unsigned int)(1 << kfd->device_info->max_pasid_bits), 94 iommu_info.max_pasids); 95 96 if (!kfd_set_pasid_limit(pasid_limit)) { 97 dev_err(kfd_device, "error setting pasid limit\n"); 98 return -EBUSY; 99 } 100 101 return 0; 102 } 103 104 /** kfd_iommu_bind_process_to_device - Have the IOMMU bind a process 105 * 106 * Binds the given process to the given device using its PASID. This 107 * enables IOMMUv2 address translation for the process on the device. 108 * 109 * This function assumes that the process mutex is held. 110 */ 111 int kfd_iommu_bind_process_to_device(struct kfd_process_device *pdd) 112 { 113 struct kfd_dev *dev = pdd->dev; 114 struct kfd_process *p = pdd->process; 115 int err; 116 117 if (!dev->device_info->needs_iommu_device || pdd->bound == PDD_BOUND) 118 return 0; 119 120 if (unlikely(pdd->bound == PDD_BOUND_SUSPENDED)) { 121 pr_err("Binding PDD_BOUND_SUSPENDED pdd is unexpected!\n"); 122 return -EINVAL; 123 } 124 125 err = amd_iommu_bind_pasid(dev->pdev, p->pasid, p->lead_thread); 126 if (!err) 127 pdd->bound = PDD_BOUND; 128 129 return err; 130 } 131 132 /** kfd_iommu_unbind_process - Unbind process from all devices 133 * 134 * This removes all IOMMU device bindings of the process. To be used 135 * before process termination. 136 */ 137 void kfd_iommu_unbind_process(struct kfd_process *p) 138 { 139 struct kfd_process_device *pdd; 140 141 list_for_each_entry(pdd, &p->per_device_data, per_device_list) 142 if (pdd->bound == PDD_BOUND) 143 amd_iommu_unbind_pasid(pdd->dev->pdev, p->pasid); 144 } 145 146 /* Callback for process shutdown invoked by the IOMMU driver */ 147 static void iommu_pasid_shutdown_callback(struct pci_dev *pdev, int pasid) 148 { 149 struct kfd_dev *dev = kfd_device_by_pci_dev(pdev); 150 struct kfd_process *p; 151 struct kfd_process_device *pdd; 152 153 if (!dev) 154 return; 155 156 /* 157 * Look for the process that matches the pasid. If there is no such 158 * process, we either released it in amdkfd's own notifier, or there 159 * is a bug. Unfortunately, there is no way to tell... 160 */ 161 p = kfd_lookup_process_by_pasid(pasid); 162 if (!p) 163 return; 164 165 pr_debug("Unbinding process 0x%x from IOMMU\n", pasid); 166 167 mutex_lock(kfd_get_dbgmgr_mutex()); 168 169 if (dev->dbgmgr && dev->dbgmgr->pasid == p->pasid) { 170 if (!kfd_dbgmgr_unregister(dev->dbgmgr, p)) { 171 kfd_dbgmgr_destroy(dev->dbgmgr); 172 dev->dbgmgr = NULL; 173 } 174 } 175 176 mutex_unlock(kfd_get_dbgmgr_mutex()); 177 178 mutex_lock(&p->mutex); 179 180 pdd = kfd_get_process_device_data(dev, p); 181 if (pdd) 182 /* For GPU relying on IOMMU, we need to dequeue here 183 * when PASID is still bound. 184 */ 185 kfd_process_dequeue_from_device(pdd); 186 187 mutex_unlock(&p->mutex); 188 189 kfd_unref_process(p); 190 } 191 192 /* This function called by IOMMU driver on PPR failure */ 193 static int iommu_invalid_ppr_cb(struct pci_dev *pdev, int pasid, 194 unsigned long address, u16 flags) 195 { 196 struct kfd_dev *dev; 197 198 dev_warn_ratelimited(kfd_device, 199 "Invalid PPR device %x:%x.%x pasid 0x%x address 0x%lX flags 0x%X", 200 PCI_BUS_NUM(pdev->devfn), 201 PCI_SLOT(pdev->devfn), 202 PCI_FUNC(pdev->devfn), 203 pasid, 204 address, 205 flags); 206 207 dev = kfd_device_by_pci_dev(pdev); 208 if (!WARN_ON(!dev)) 209 kfd_signal_iommu_event(dev, pasid, address, 210 flags & PPR_FAULT_WRITE, flags & PPR_FAULT_EXEC); 211 212 return AMD_IOMMU_INV_PRI_RSP_INVALID; 213 } 214 215 /* 216 * Bind processes do the device that have been temporarily unbound 217 * (PDD_BOUND_SUSPENDED) in kfd_unbind_processes_from_device. 218 */ 219 static int kfd_bind_processes_to_device(struct kfd_dev *kfd) 220 { 221 struct kfd_process_device *pdd; 222 struct kfd_process *p; 223 unsigned int temp; 224 int err = 0; 225 226 int idx = srcu_read_lock(&kfd_processes_srcu); 227 228 hash_for_each_rcu(kfd_processes_table, temp, p, kfd_processes) { 229 mutex_lock(&p->mutex); 230 pdd = kfd_get_process_device_data(kfd, p); 231 232 if (WARN_ON(!pdd) || pdd->bound != PDD_BOUND_SUSPENDED) { 233 mutex_unlock(&p->mutex); 234 continue; 235 } 236 237 err = amd_iommu_bind_pasid(kfd->pdev, p->pasid, 238 p->lead_thread); 239 if (err < 0) { 240 pr_err("Unexpected pasid 0x%x binding failure\n", 241 p->pasid); 242 mutex_unlock(&p->mutex); 243 break; 244 } 245 246 pdd->bound = PDD_BOUND; 247 mutex_unlock(&p->mutex); 248 } 249 250 srcu_read_unlock(&kfd_processes_srcu, idx); 251 252 return err; 253 } 254 255 /* 256 * Mark currently bound processes as PDD_BOUND_SUSPENDED. These 257 * processes will be restored to PDD_BOUND state in 258 * kfd_bind_processes_to_device. 259 */ 260 static void kfd_unbind_processes_from_device(struct kfd_dev *kfd) 261 { 262 struct kfd_process_device *pdd; 263 struct kfd_process *p; 264 unsigned int temp; 265 266 int idx = srcu_read_lock(&kfd_processes_srcu); 267 268 hash_for_each_rcu(kfd_processes_table, temp, p, kfd_processes) { 269 mutex_lock(&p->mutex); 270 pdd = kfd_get_process_device_data(kfd, p); 271 272 if (WARN_ON(!pdd)) { 273 mutex_unlock(&p->mutex); 274 continue; 275 } 276 277 if (pdd->bound == PDD_BOUND) 278 pdd->bound = PDD_BOUND_SUSPENDED; 279 mutex_unlock(&p->mutex); 280 } 281 282 srcu_read_unlock(&kfd_processes_srcu, idx); 283 } 284 285 /** kfd_iommu_suspend - Prepare IOMMU for suspend 286 * 287 * This unbinds processes from the device and disables the IOMMU for 288 * the device. 289 */ 290 void kfd_iommu_suspend(struct kfd_dev *kfd) 291 { 292 if (!kfd->device_info->needs_iommu_device) 293 return; 294 295 kfd_unbind_processes_from_device(kfd); 296 297 amd_iommu_set_invalidate_ctx_cb(kfd->pdev, NULL); 298 amd_iommu_set_invalid_ppr_cb(kfd->pdev, NULL); 299 amd_iommu_free_device(kfd->pdev); 300 } 301 302 /** kfd_iommu_resume - Restore IOMMU after resume 303 * 304 * This reinitializes the IOMMU for the device and re-binds previously 305 * suspended processes to the device. 306 */ 307 int kfd_iommu_resume(struct kfd_dev *kfd) 308 { 309 unsigned int pasid_limit; 310 int err; 311 312 if (!kfd->device_info->needs_iommu_device) 313 return 0; 314 315 pasid_limit = kfd_get_pasid_limit(); 316 317 err = amd_iommu_init_device(kfd->pdev, pasid_limit); 318 if (err) 319 return -ENXIO; 320 321 amd_iommu_set_invalidate_ctx_cb(kfd->pdev, 322 iommu_pasid_shutdown_callback); 323 amd_iommu_set_invalid_ppr_cb(kfd->pdev, 324 iommu_invalid_ppr_cb); 325 326 err = kfd_bind_processes_to_device(kfd); 327 if (err) { 328 amd_iommu_set_invalidate_ctx_cb(kfd->pdev, NULL); 329 amd_iommu_set_invalid_ppr_cb(kfd->pdev, NULL); 330 amd_iommu_free_device(kfd->pdev); 331 return err; 332 } 333 334 return 0; 335 } 336 337 extern bool amd_iommu_pc_supported(void); 338 extern u8 amd_iommu_pc_get_max_banks(u16 devid); 339 extern u8 amd_iommu_pc_get_max_counters(u16 devid); 340 341 /** kfd_iommu_add_perf_counters - Add IOMMU performance counters to topology 342 */ 343 int kfd_iommu_add_perf_counters(struct kfd_topology_device *kdev) 344 { 345 struct kfd_perf_properties *props; 346 347 if (!(kdev->node_props.capability & HSA_CAP_ATS_PRESENT)) 348 return 0; 349 350 if (!amd_iommu_pc_supported()) 351 return 0; 352 353 props = kfd_alloc_struct(props); 354 if (!props) 355 return -ENOMEM; 356 strcpy(props->block_name, "iommu"); 357 props->max_concurrent = amd_iommu_pc_get_max_banks(0) * 358 amd_iommu_pc_get_max_counters(0); /* assume one iommu */ 359 list_add_tail(&props->list, &kdev->perf_props); 360 361 return 0; 362 } 363