Consistently use round_page(x) rather than roundup(x, PAGE_SIZE). There is
no functional change.
This commit is contained in:
parent
2787e39aaa
commit
00a54dfb1c
Notes:
svn2git
2020-12-20 02:59:44 +00:00
svn path=/head/; revision=246855
@ -122,7 +122,7 @@ table_map(vm_paddr_t pa, int offset, vm_offset_t length)
|
||||
void *data;
|
||||
|
||||
off = pa & PAGE_MASK;
|
||||
length = roundup(length + off, PAGE_SIZE);
|
||||
length = round_page(length + off);
|
||||
pa = pa & PG_FRAME;
|
||||
va = (vm_offset_t)pmap_kenter_temporary(pa, offset) +
|
||||
(offset * PAGE_SIZE);
|
||||
@ -146,7 +146,7 @@ table_unmap(void *data, vm_offset_t length)
|
||||
|
||||
va = (vm_offset_t)data;
|
||||
off = va & PAGE_MASK;
|
||||
length = roundup(length + off, PAGE_SIZE);
|
||||
length = round_page(length + off);
|
||||
va &= ~PAGE_MASK;
|
||||
while (length > 0) {
|
||||
pmap_kremove(va);
|
||||
|
@ -5015,7 +5015,7 @@ pmap_mapdev_attr(vm_paddr_t pa, vm_size_t size, int mode)
|
||||
return ((void *)va);
|
||||
}
|
||||
offset = pa & PAGE_MASK;
|
||||
size = roundup(offset + size, PAGE_SIZE);
|
||||
size = round_page(offset + size);
|
||||
va = kmem_alloc_nofault(kernel_map, size);
|
||||
if (!va)
|
||||
panic("pmap_mapdev: Couldn't alloc kernel virtual memory");
|
||||
@ -5051,7 +5051,7 @@ pmap_unmapdev(vm_offset_t va, vm_size_t size)
|
||||
return;
|
||||
base = trunc_page(va);
|
||||
offset = va & PAGE_MASK;
|
||||
size = roundup(offset + size, PAGE_SIZE);
|
||||
size = round_page(offset + size);
|
||||
kmem_free(kernel_map, base, size);
|
||||
}
|
||||
|
||||
@ -5173,7 +5173,7 @@ pmap_change_attr_locked(vm_offset_t va, vm_size_t size, int mode)
|
||||
PMAP_LOCK_ASSERT(kernel_pmap, MA_OWNED);
|
||||
base = trunc_page(va);
|
||||
offset = va & PAGE_MASK;
|
||||
size = roundup(offset + size, PAGE_SIZE);
|
||||
size = round_page(offset + size);
|
||||
|
||||
/*
|
||||
* Only supported on kernel virtual addresses, including the direct
|
||||
|
@ -139,7 +139,7 @@ table_map(vm_paddr_t pa, int offset, vm_offset_t length)
|
||||
void *data;
|
||||
|
||||
off = pa & PAGE_MASK;
|
||||
length = roundup(length + off, PAGE_SIZE);
|
||||
length = round_page(length + off);
|
||||
pa = pa & PG_FRAME;
|
||||
va = (vm_offset_t)pmap_kenter_temporary(pa, offset) +
|
||||
(offset * PAGE_SIZE);
|
||||
@ -163,7 +163,7 @@ table_unmap(void *data, vm_offset_t length)
|
||||
|
||||
va = (vm_offset_t)data;
|
||||
off = va & PAGE_MASK;
|
||||
length = roundup(length + off, PAGE_SIZE);
|
||||
length = round_page(length + off);
|
||||
va &= ~PAGE_MASK;
|
||||
while (length > 0) {
|
||||
pmap_kremove(va);
|
||||
|
@ -4989,7 +4989,7 @@ pmap_mapdev_attr(vm_paddr_t pa, vm_size_t size, int mode)
|
||||
vm_size_t tmpsize;
|
||||
|
||||
offset = pa & PAGE_MASK;
|
||||
size = roundup(offset + size, PAGE_SIZE);
|
||||
size = round_page(offset + size);
|
||||
pa = pa & PG_FRAME;
|
||||
|
||||
if (pa < KERNLOAD && pa + size <= KERNLOAD)
|
||||
@ -5029,7 +5029,7 @@ pmap_unmapdev(vm_offset_t va, vm_size_t size)
|
||||
return;
|
||||
base = trunc_page(va);
|
||||
offset = va & PAGE_MASK;
|
||||
size = roundup(offset + size, PAGE_SIZE);
|
||||
size = round_page(offset + size);
|
||||
kmem_free(kernel_map, base, size);
|
||||
}
|
||||
|
||||
@ -5120,7 +5120,7 @@ pmap_change_attr(vm_offset_t va, vm_size_t size, int mode)
|
||||
|
||||
base = trunc_page(va);
|
||||
offset = va & PAGE_MASK;
|
||||
size = roundup(offset + size, PAGE_SIZE);
|
||||
size = round_page(offset + size);
|
||||
|
||||
/*
|
||||
* Only supported on kernel virtual addresses above the recursive map.
|
||||
|
@ -3973,7 +3973,7 @@ pmap_mapdev_attr(vm_paddr_t pa, vm_size_t size, int mode)
|
||||
vm_size_t tmpsize;
|
||||
|
||||
offset = pa & PAGE_MASK;
|
||||
size = roundup(offset + size, PAGE_SIZE);
|
||||
size = round_page(offset + size);
|
||||
pa = pa & PG_FRAME;
|
||||
|
||||
if (pa < KERNLOAD && pa + size <= KERNLOAD)
|
||||
@ -4013,7 +4013,7 @@ pmap_unmapdev(vm_offset_t va, vm_size_t size)
|
||||
return;
|
||||
base = trunc_page(va);
|
||||
offset = va & PAGE_MASK;
|
||||
size = roundup(offset + size, PAGE_SIZE);
|
||||
size = round_page(offset + size);
|
||||
kmem_free(kernel_map, base, size);
|
||||
}
|
||||
|
||||
@ -4105,7 +4105,7 @@ pmap_change_attr(vm_offset_t va, vm_size_t size, int mode)
|
||||
|
||||
base = trunc_page(va);
|
||||
offset = va & PAGE_MASK;
|
||||
size = roundup(offset + size, PAGE_SIZE);
|
||||
size = round_page(offset + size);
|
||||
|
||||
/* Only supported on kernel virtual addresses. */
|
||||
if (base <= VM_MAXUSER_ADDRESS)
|
||||
|
Loading…
Reference in New Issue
Block a user