static struct kmem_cache *iommu_domain_cache;
static struct kmem_cache *iommu_devinfo_cache;
+static struct dmar_domain* get_iommu_domain(struct intel_iommu *iommu, u16 did)
+{
+ return iommu->domains[did];
+}
+
+static void set_iommu_domain(struct intel_iommu *iommu, u16 did,
+ struct dmar_domain *domain)
+{
+ iommu->domains[did] = domain;
+}
+
static inline void *alloc_pgtable_page(int node)
{
struct page *page;
* flush. However, device IOTLB doesn't need to be flushed in this case.
*/
if (!cap_caching_mode(iommu->cap) || !map)
- iommu_flush_dev_iotlb(iommu->domains[did], addr, mask);
+ iommu_flush_dev_iotlb(get_iommu_domain(iommu, did),
+ addr, mask);
}
static void iommu_disable_protect_mem_regions(struct intel_iommu *iommu)
if (i == 0)
continue;
- domain = iommu->domains[i];
+ domain = get_iommu_domain(iommu, i);
clear_bit(i, iommu->domain_ids);
if (domain_detach_iommu(domain, iommu) == 0 &&
!domain_type_is_vm(domain))
num = find_first_zero_bit(iommu->domain_ids, ndomains);
if (num < ndomains) {
set_bit(num, iommu->domain_ids);
- iommu->domains[num] = domain;
+ set_iommu_domain(iommu, num, domain);
domain->iommu_did[iommu->seq_id] = num;
} else {
num = -ENOSPC;
return;
clear_bit(num, iommu->domain_ids);
- iommu->domains[num] = NULL;
+ set_iommu_domain(iommu, num, NULL);
spin_unlock_irqrestore(&iommu->lock, flags);
}
*/
ndomains = cap_ndoms(iommu->cap);
for_each_set_bit(num, iommu->domain_ids, ndomains) {
- if (iommu->domains[num] == dmar_domain)
+ if (get_iommu_domain(iommu, num) == dmar_domain)
iommu_flush_iotlb_psi(iommu, num, start_pfn,
npages, !freelist, 0);
}