{
unsigned long vaddr;
pgd_t *pgd_base = swapper_pg_dir;
- int i;
#if CONFIG_X86_PAE
+ int i;
/* Init entries of the first-level page table to the zero page */
for (i = 0; i < PTRS_PER_PGD; i++)
set_pgd(pgd_base + i, __pgd(__pa(empty_zero_page) | _PAGE_PRESENT));
* Associate a virtual page frame with a given physical page frame
* and protection flags for that frame.
*/
-static void set_pte_phys (unsigned long vaddr, unsigned long phys, pgprot_t flags)
+static void set_pte_pfn(unsigned long vaddr, unsigned long pfn, pgprot_t flags)
{
pgd_t *pgd;
pmd_t *pmd;
return;
}
pte = pte_offset_kernel(pmd, vaddr);
- /* <phys,flags> stored as-is, to permit clearing entries */
- set_pte(pte, pfn_pte(phys >> PAGE_SHIFT, flags));
+ /* <pfn,flags> stored as-is, to permit clearing entries */
+ set_pte(pte, pfn_pte(pfn, flags));
/*
* It's enough to flush this one mapping.
BUG();
return;
}
- set_pte_phys(address, phys, flags);
+ set_pte_pfn(address, phys >> PAGE_SHIFT, flags);
}
pte_t *pte_alloc_one_kernel(struct mm_struct *mm, unsigned long address)