Consistently use round_page(x) rather than roundup(x, PAGE_SIZE). There is

no functional change.
This commit is contained in:
Jung-uk Kim 2013-02-15 22:43:08 +00:00
parent 2787e39aaa
commit 00a54dfb1c
Notes: svn2git 2020-12-20 02:59:44 +00:00
svn path=/head/; revision=246855
5 changed files with 13 additions and 13 deletions

View File

@ -122,7 +122,7 @@ table_map(vm_paddr_t pa, int offset, vm_offset_t length)
void *data;
off = pa & PAGE_MASK;
length = roundup(length + off, PAGE_SIZE);
length = round_page(length + off);
pa = pa & PG_FRAME;
va = (vm_offset_t)pmap_kenter_temporary(pa, offset) +
(offset * PAGE_SIZE);
@ -146,7 +146,7 @@ table_unmap(void *data, vm_offset_t length)
va = (vm_offset_t)data;
off = va & PAGE_MASK;
length = roundup(length + off, PAGE_SIZE);
length = round_page(length + off);
va &= ~PAGE_MASK;
while (length > 0) {
pmap_kremove(va);

View File

@ -5015,7 +5015,7 @@ pmap_mapdev_attr(vm_paddr_t pa, vm_size_t size, int mode)
return ((void *)va);
}
offset = pa & PAGE_MASK;
size = roundup(offset + size, PAGE_SIZE);
size = round_page(offset + size);
va = kmem_alloc_nofault(kernel_map, size);
if (!va)
panic("pmap_mapdev: Couldn't alloc kernel virtual memory");
@ -5051,7 +5051,7 @@ pmap_unmapdev(vm_offset_t va, vm_size_t size)
return;
base = trunc_page(va);
offset = va & PAGE_MASK;
size = roundup(offset + size, PAGE_SIZE);
size = round_page(offset + size);
kmem_free(kernel_map, base, size);
}
@ -5173,7 +5173,7 @@ pmap_change_attr_locked(vm_offset_t va, vm_size_t size, int mode)
PMAP_LOCK_ASSERT(kernel_pmap, MA_OWNED);
base = trunc_page(va);
offset = va & PAGE_MASK;
size = roundup(offset + size, PAGE_SIZE);
size = round_page(offset + size);
/*
* Only supported on kernel virtual addresses, including the direct

View File

@ -139,7 +139,7 @@ table_map(vm_paddr_t pa, int offset, vm_offset_t length)
void *data;
off = pa & PAGE_MASK;
length = roundup(length + off, PAGE_SIZE);
length = round_page(length + off);
pa = pa & PG_FRAME;
va = (vm_offset_t)pmap_kenter_temporary(pa, offset) +
(offset * PAGE_SIZE);
@ -163,7 +163,7 @@ table_unmap(void *data, vm_offset_t length)
va = (vm_offset_t)data;
off = va & PAGE_MASK;
length = roundup(length + off, PAGE_SIZE);
length = round_page(length + off);
va &= ~PAGE_MASK;
while (length > 0) {
pmap_kremove(va);

View File

@ -4989,7 +4989,7 @@ pmap_mapdev_attr(vm_paddr_t pa, vm_size_t size, int mode)
vm_size_t tmpsize;
offset = pa & PAGE_MASK;
size = roundup(offset + size, PAGE_SIZE);
size = round_page(offset + size);
pa = pa & PG_FRAME;
if (pa < KERNLOAD && pa + size <= KERNLOAD)
@ -5029,7 +5029,7 @@ pmap_unmapdev(vm_offset_t va, vm_size_t size)
return;
base = trunc_page(va);
offset = va & PAGE_MASK;
size = roundup(offset + size, PAGE_SIZE);
size = round_page(offset + size);
kmem_free(kernel_map, base, size);
}
@ -5120,7 +5120,7 @@ pmap_change_attr(vm_offset_t va, vm_size_t size, int mode)
base = trunc_page(va);
offset = va & PAGE_MASK;
size = roundup(offset + size, PAGE_SIZE);
size = round_page(offset + size);
/*
* Only supported on kernel virtual addresses above the recursive map.

View File

@ -3973,7 +3973,7 @@ pmap_mapdev_attr(vm_paddr_t pa, vm_size_t size, int mode)
vm_size_t tmpsize;
offset = pa & PAGE_MASK;
size = roundup(offset + size, PAGE_SIZE);
size = round_page(offset + size);
pa = pa & PG_FRAME;
if (pa < KERNLOAD && pa + size <= KERNLOAD)
@ -4013,7 +4013,7 @@ pmap_unmapdev(vm_offset_t va, vm_size_t size)
return;
base = trunc_page(va);
offset = va & PAGE_MASK;
size = roundup(offset + size, PAGE_SIZE);
size = round_page(offset + size);
kmem_free(kernel_map, base, size);
}
@ -4105,7 +4105,7 @@ pmap_change_attr(vm_offset_t va, vm_size_t size, int mode)
base = trunc_page(va);
offset = va & PAGE_MASK;
size = roundup(offset + size, PAGE_SIZE);
size = round_page(offset + size);
/* Only supported on kernel virtual addresses. */
if (base <= VM_MAXUSER_ADDRESS)