#include <asm/iommu.h>
#include <asm/tce.h>
-static unsigned long kvmppc_tce_pages(unsigned long window_size)
+static unsigned long kvmppc_tce_pages(unsigned long iommu_pages)
{
- return ALIGN((window_size >> IOMMU_PAGE_SHIFT_4K)
- * sizeof(u64), PAGE_SIZE) / PAGE_SIZE;
+ return ALIGN(iommu_pages * sizeof(u64), PAGE_SIZE) / PAGE_SIZE;
}
static unsigned long kvmppc_stt_pages(unsigned long tce_pages)
{
struct kvmppc_spapr_tce_table *stt = container_of(head,
struct kvmppc_spapr_tce_table, rcu);
- int i;
- unsigned long npages = kvmppc_tce_pages(stt->window_size);
+ unsigned long i, npages = kvmppc_tce_pages(stt->size);
for (i = 0; i < npages; i++)
__free_page(stt->pages[i]);
struct kvmppc_spapr_tce_table *stt = vma->vm_file->private_data;
struct page *page;
- if (vmf->pgoff >= kvmppc_tce_pages(stt->window_size))
+ if (vmf->pgoff >= kvmppc_tce_pages(stt->size))
return VM_FAULT_SIGBUS;
page = stt->pages[vmf->pgoff];
kvm_put_kvm(stt->kvm);
kvmppc_account_memlimit(
- kvmppc_stt_pages(kvmppc_tce_pages(stt->window_size)), false);
+ kvmppc_stt_pages(kvmppc_tce_pages(stt->size)), false);
call_rcu(&stt->rcu, release_spapr_tce_table);
return 0;
struct kvm_create_spapr_tce *args)
{
struct kvmppc_spapr_tce_table *stt = NULL;
- unsigned long npages;
+ unsigned long npages, size;
int ret = -ENOMEM;
int i;
return -EBUSY;
}
- npages = kvmppc_tce_pages(args->window_size);
+ size = args->window_size >> IOMMU_PAGE_SHIFT_4K;
+ npages = kvmppc_tce_pages(size);
ret = kvmppc_account_memlimit(kvmppc_stt_pages(npages), true);
if (ret) {
stt = NULL;
goto fail;
stt->liobn = args->liobn;
- stt->window_size = args->window_size;
+ stt->page_shift = IOMMU_PAGE_SHIFT_4K;
+ stt->size = size;
stt->kvm = kvm;
for (i = 0; i < npages; i++) {
if (!stt)
return H_TOO_HARD;
- entry = ioba >> IOMMU_PAGE_SHIFT_4K;
+ entry = ioba >> stt->page_shift;
/*
* SPAPR spec says that the maximum size of the list is 512 TCEs
* so the whole table fits in 4K page
long kvmppc_ioba_validate(struct kvmppc_spapr_tce_table *stt,
unsigned long ioba, unsigned long npages)
{
- unsigned long mask = (1ULL << IOMMU_PAGE_SHIFT_4K) - 1;
- unsigned long idx = ioba >> IOMMU_PAGE_SHIFT_4K;
- unsigned long size = stt->window_size >> IOMMU_PAGE_SHIFT_4K;
+ unsigned long mask = (1ULL << stt->page_shift) - 1;
+ unsigned long idx = ioba >> stt->page_shift;
- if ((ioba & mask) || (idx + npages > size) || (idx + npages < idx))
+ if ((ioba & mask) || (idx + npages > stt->size) || (idx + npages < idx))
return H_PARAMETER;
return H_SUCCESS;
*/
long kvmppc_tce_validate(struct kvmppc_spapr_tce_table *stt, unsigned long tce)
{
- unsigned long mask =
- ~(IOMMU_PAGE_MASK_4K | TCE_PCI_WRITE | TCE_PCI_READ);
+ unsigned long page_mask = ~((1ULL << stt->page_shift) - 1);
+ unsigned long mask = ~(page_mask | TCE_PCI_WRITE | TCE_PCI_READ);
if (tce & mask)
return H_PARAMETER;
if (ret != H_SUCCESS)
return ret;
- kvmppc_tce_put(stt, ioba >> IOMMU_PAGE_SHIFT_4K, tce);
+ kvmppc_tce_put(stt, ioba >> stt->page_shift, tce);
return H_SUCCESS;
}
if (!stt)
return H_TOO_HARD;
- entry = ioba >> IOMMU_PAGE_SHIFT_4K;
+ entry = ioba >> stt->page_shift;
/*
* The spec says that the maximum size of the list is 512 TCEs
* so the whole table addressed resides in 4K page
if (tce_value & (TCE_PCI_WRITE | TCE_PCI_READ))
return H_PARAMETER;
- for (i = 0; i < npages; ++i, ioba += IOMMU_PAGE_SIZE_4K)
- kvmppc_tce_put(stt, ioba >> IOMMU_PAGE_SHIFT_4K, tce_value);
+ for (i = 0; i < npages; ++i, ioba += (1ULL << stt->page_shift))
+ kvmppc_tce_put(stt, ioba >> stt->page_shift, tce_value);
return H_SUCCESS;
}
if (ret != H_SUCCESS)
return ret;
- idx = ioba >> IOMMU_PAGE_SHIFT_4K;
+ idx = ioba >> stt->page_shift;
page = stt->pages[idx / TCES_PER_PAGE];
tbl = (u64 *)page_address(page);