Loading drivers/iommu/ipmmu-vmsa.c +9 −57 Original line number Diff line number Diff line Loading @@ -215,7 +215,6 @@ static LIST_HEAD(ipmmu_devices); #define IPMMU_PTRS_PER_PTE 512 #define IPMMU_PTRS_PER_PMD 512 #define IPMMU_PTRS_PER_PGD 4 #define IPMMU_PTRS_PER_PUD 1 /* ----------------------------------------------------------------------------- * Read/Write Access Loading Loading @@ -465,6 +464,8 @@ static irqreturn_t ipmmu_irq(int irq, void *dev) * Page Table Management */ #define pud_pgtable(pud) pfn_to_page(__phys_to_pfn(pud_val(pud) & PHYS_MASK)) static void ipmmu_free_ptes(pmd_t *pmd) { pgtable_t table = pmd_pgtable(*pmd); Loading @@ -473,10 +474,10 @@ static void ipmmu_free_ptes(pmd_t *pmd) static void ipmmu_free_pmds(pud_t *pud) { pmd_t *pmd, *pmd_base = pmd_offset(pud, 0); pmd_t *pmd = pmd_offset(pud, 0); pgtable_t table; unsigned int i; pmd = pmd_base; for (i = 0; i < IPMMU_PTRS_PER_PMD; ++i) { if (pmd_none(*pmd)) continue; Loading @@ -485,24 +486,8 @@ static void ipmmu_free_pmds(pud_t *pud) pmd++; } pmd_free(NULL, pmd_base); } static void ipmmu_free_puds(pgd_t *pgd) { pud_t *pud, *pud_base = pud_offset(pgd, 0); unsigned int i; pud = pud_base; for (i = 0; i < IPMMU_PTRS_PER_PUD; ++i) { if (pud_none(*pud)) continue; ipmmu_free_pmds(pud); pud++; } pud_free(NULL, pud_base); table = pud_pgtable(*pud); __free_page(table); } static void ipmmu_free_pgtables(struct ipmmu_vmsa_domain *domain) Loading @@ -520,7 +505,7 @@ static void ipmmu_free_pgtables(struct ipmmu_vmsa_domain *domain) for (i = 0; i < IPMMU_PTRS_PER_PGD; ++i) { if (pgd_none(*pgd)) continue; ipmmu_free_puds(pgd); ipmmu_free_pmds((pud_t *)pgd); pgd++; } Loading Loading @@ -624,7 +609,6 @@ static int ipmmu_alloc_init_pmd(struct ipmmu_vmsa_device *mmu, pud_t *pud, pmd_t *pmd; int ret; #ifndef __PAGETABLE_PMD_FOLDED if (pud_none(*pud)) { pmd = (pmd_t *)get_zeroed_page(GFP_ATOMIC); if (!pmd) Loading @@ -636,7 +620,6 @@ static int ipmmu_alloc_init_pmd(struct ipmmu_vmsa_device *mmu, pud_t *pud, pmd += pmd_index(addr); } else #endif pmd = pmd_offset(pud, addr); do { Loading @@ -648,38 +631,6 @@ static int ipmmu_alloc_init_pmd(struct ipmmu_vmsa_device *mmu, pud_t *pud, return ret; } static int ipmmu_alloc_init_pud(struct ipmmu_vmsa_device *mmu, pgd_t *pgd, unsigned long addr, unsigned long end, phys_addr_t phys, int prot) { unsigned long next; pud_t *pud; int ret; #ifndef __PAGETABLE_PUD_FOLDED if (pgd_none(*pgd)) { pud = (pud_t *)get_zeroed_page(GFP_ATOMIC); if (!pud) return -ENOMEM; ipmmu_flush_pgtable(mmu, pud, PAGE_SIZE); *pgd = __pgd(__pa(pud) | PMD_NSTABLE | PMD_TYPE_TABLE); ipmmu_flush_pgtable(mmu, pgd, sizeof(*pgd)); pud += pud_index(addr); } else #endif pud = pud_offset(pgd, addr); do { next = pud_addr_end(addr, end); ret = ipmmu_alloc_init_pmd(mmu, pud, addr, next, phys, prot); phys += next - addr; } while (pud++, addr = next, addr < end); return ret; } static int ipmmu_handle_mapping(struct ipmmu_vmsa_domain *domain, unsigned long iova, phys_addr_t paddr, size_t size, int prot) Loading Loading @@ -707,7 +658,8 @@ static int ipmmu_handle_mapping(struct ipmmu_vmsa_domain *domain, do { unsigned long next = pgd_addr_end(iova, end); ret = ipmmu_alloc_init_pud(mmu, pgd, iova, next, paddr, prot); ret = ipmmu_alloc_init_pmd(mmu, (pud_t *)pgd, iova, next, paddr, prot); if (ret) break; Loading Loading
drivers/iommu/ipmmu-vmsa.c +9 −57 Original line number Diff line number Diff line Loading @@ -215,7 +215,6 @@ static LIST_HEAD(ipmmu_devices); #define IPMMU_PTRS_PER_PTE 512 #define IPMMU_PTRS_PER_PMD 512 #define IPMMU_PTRS_PER_PGD 4 #define IPMMU_PTRS_PER_PUD 1 /* ----------------------------------------------------------------------------- * Read/Write Access Loading Loading @@ -465,6 +464,8 @@ static irqreturn_t ipmmu_irq(int irq, void *dev) * Page Table Management */ #define pud_pgtable(pud) pfn_to_page(__phys_to_pfn(pud_val(pud) & PHYS_MASK)) static void ipmmu_free_ptes(pmd_t *pmd) { pgtable_t table = pmd_pgtable(*pmd); Loading @@ -473,10 +474,10 @@ static void ipmmu_free_ptes(pmd_t *pmd) static void ipmmu_free_pmds(pud_t *pud) { pmd_t *pmd, *pmd_base = pmd_offset(pud, 0); pmd_t *pmd = pmd_offset(pud, 0); pgtable_t table; unsigned int i; pmd = pmd_base; for (i = 0; i < IPMMU_PTRS_PER_PMD; ++i) { if (pmd_none(*pmd)) continue; Loading @@ -485,24 +486,8 @@ static void ipmmu_free_pmds(pud_t *pud) pmd++; } pmd_free(NULL, pmd_base); } static void ipmmu_free_puds(pgd_t *pgd) { pud_t *pud, *pud_base = pud_offset(pgd, 0); unsigned int i; pud = pud_base; for (i = 0; i < IPMMU_PTRS_PER_PUD; ++i) { if (pud_none(*pud)) continue; ipmmu_free_pmds(pud); pud++; } pud_free(NULL, pud_base); table = pud_pgtable(*pud); __free_page(table); } static void ipmmu_free_pgtables(struct ipmmu_vmsa_domain *domain) Loading @@ -520,7 +505,7 @@ static void ipmmu_free_pgtables(struct ipmmu_vmsa_domain *domain) for (i = 0; i < IPMMU_PTRS_PER_PGD; ++i) { if (pgd_none(*pgd)) continue; ipmmu_free_puds(pgd); ipmmu_free_pmds((pud_t *)pgd); pgd++; } Loading Loading @@ -624,7 +609,6 @@ static int ipmmu_alloc_init_pmd(struct ipmmu_vmsa_device *mmu, pud_t *pud, pmd_t *pmd; int ret; #ifndef __PAGETABLE_PMD_FOLDED if (pud_none(*pud)) { pmd = (pmd_t *)get_zeroed_page(GFP_ATOMIC); if (!pmd) Loading @@ -636,7 +620,6 @@ static int ipmmu_alloc_init_pmd(struct ipmmu_vmsa_device *mmu, pud_t *pud, pmd += pmd_index(addr); } else #endif pmd = pmd_offset(pud, addr); do { Loading @@ -648,38 +631,6 @@ static int ipmmu_alloc_init_pmd(struct ipmmu_vmsa_device *mmu, pud_t *pud, return ret; } static int ipmmu_alloc_init_pud(struct ipmmu_vmsa_device *mmu, pgd_t *pgd, unsigned long addr, unsigned long end, phys_addr_t phys, int prot) { unsigned long next; pud_t *pud; int ret; #ifndef __PAGETABLE_PUD_FOLDED if (pgd_none(*pgd)) { pud = (pud_t *)get_zeroed_page(GFP_ATOMIC); if (!pud) return -ENOMEM; ipmmu_flush_pgtable(mmu, pud, PAGE_SIZE); *pgd = __pgd(__pa(pud) | PMD_NSTABLE | PMD_TYPE_TABLE); ipmmu_flush_pgtable(mmu, pgd, sizeof(*pgd)); pud += pud_index(addr); } else #endif pud = pud_offset(pgd, addr); do { next = pud_addr_end(addr, end); ret = ipmmu_alloc_init_pmd(mmu, pud, addr, next, phys, prot); phys += next - addr; } while (pud++, addr = next, addr < end); return ret; } static int ipmmu_handle_mapping(struct ipmmu_vmsa_domain *domain, unsigned long iova, phys_addr_t paddr, size_t size, int prot) Loading Loading @@ -707,7 +658,8 @@ static int ipmmu_handle_mapping(struct ipmmu_vmsa_domain *domain, do { unsigned long next = pgd_addr_end(iova, end); ret = ipmmu_alloc_init_pud(mmu, pgd, iova, next, paddr, prot); ret = ipmmu_alloc_init_pmd(mmu, (pud_t *)pgd, iova, next, paddr, prot); if (ret) break; Loading