Eliminate kmem_malloc()'s unused arena parameter. (The arena parameter
became unused in FreeBSD 12.x as a side-effect of the NUMA-related changes.) Reviewed by: kib, markj Discussed with: jeff, re@ Differential Revision: https://reviews.freebsd.org/D16825
This commit is contained in:
parent
b1a90834bb
commit
83a90bffd8
Notes:
svn2git
2020-12-20 02:59:44 +00:00
svn path=/head/; revision=338143
@ -402,18 +402,14 @@ native_start_all_aps(void)
|
|||||||
apic_id = cpu_apic_ids[cpu];
|
apic_id = cpu_apic_ids[cpu];
|
||||||
|
|
||||||
/* allocate and set up an idle stack data page */
|
/* allocate and set up an idle stack data page */
|
||||||
bootstacks[cpu] = (void *)kmem_malloc(kernel_arena,
|
bootstacks[cpu] = (void *)kmem_malloc(kstack_pages * PAGE_SIZE,
|
||||||
kstack_pages * PAGE_SIZE, M_WAITOK | M_ZERO);
|
|
||||||
doublefault_stack = (char *)kmem_malloc(kernel_arena,
|
|
||||||
PAGE_SIZE, M_WAITOK | M_ZERO);
|
|
||||||
mce_stack = (char *)kmem_malloc(kernel_arena, PAGE_SIZE,
|
|
||||||
M_WAITOK | M_ZERO);
|
|
||||||
nmi_stack = (char *)kmem_malloc(kernel_arena, PAGE_SIZE,
|
|
||||||
M_WAITOK | M_ZERO);
|
|
||||||
dbg_stack = (char *)kmem_malloc(kernel_arena, PAGE_SIZE,
|
|
||||||
M_WAITOK | M_ZERO);
|
|
||||||
dpcpu = (void *)kmem_malloc(kernel_arena, DPCPU_SIZE,
|
|
||||||
M_WAITOK | M_ZERO);
|
M_WAITOK | M_ZERO);
|
||||||
|
doublefault_stack = (char *)kmem_malloc(PAGE_SIZE, M_WAITOK |
|
||||||
|
M_ZERO);
|
||||||
|
mce_stack = (char *)kmem_malloc(PAGE_SIZE, M_WAITOK | M_ZERO);
|
||||||
|
nmi_stack = (char *)kmem_malloc(PAGE_SIZE, M_WAITOK | M_ZERO);
|
||||||
|
dbg_stack = (char *)kmem_malloc(PAGE_SIZE, M_WAITOK | M_ZERO);
|
||||||
|
dpcpu = (void *)kmem_malloc(DPCPU_SIZE, M_WAITOK | M_ZERO);
|
||||||
|
|
||||||
bootSTK = (char *)bootstacks[cpu] + kstack_pages * PAGE_SIZE - 8;
|
bootSTK = (char *)bootstacks[cpu] + kstack_pages * PAGE_SIZE - 8;
|
||||||
bootAP = cpu;
|
bootAP = cpu;
|
||||||
|
@ -1412,8 +1412,7 @@ pmap_init(void)
|
|||||||
*/
|
*/
|
||||||
s = (vm_size_t)(pv_npg * sizeof(struct md_page));
|
s = (vm_size_t)(pv_npg * sizeof(struct md_page));
|
||||||
s = round_page(s);
|
s = round_page(s);
|
||||||
pv_table = (struct md_page *)kmem_malloc(kernel_arena, s,
|
pv_table = (struct md_page *)kmem_malloc(s, M_WAITOK | M_ZERO);
|
||||||
M_WAITOK | M_ZERO);
|
|
||||||
for (i = 0; i < pv_npg; i++)
|
for (i = 0; i < pv_npg; i++)
|
||||||
TAILQ_INIT(&pv_table[i].pv_list);
|
TAILQ_INIT(&pv_table[i].pv_list);
|
||||||
TAILQ_INIT(&pv_dummy.pv_list);
|
TAILQ_INIT(&pv_dummy.pv_list);
|
||||||
|
@ -361,8 +361,8 @@ amd64_set_ioperm(td, uap)
|
|||||||
*/
|
*/
|
||||||
pcb = td->td_pcb;
|
pcb = td->td_pcb;
|
||||||
if (pcb->pcb_tssp == NULL) {
|
if (pcb->pcb_tssp == NULL) {
|
||||||
tssp = (struct amd64tss *)kmem_malloc(kernel_arena,
|
tssp = (struct amd64tss *)kmem_malloc(ctob(IOPAGES + 1),
|
||||||
ctob(IOPAGES + 1), M_WAITOK);
|
M_WAITOK);
|
||||||
pmap_pti_add_kva((vm_offset_t)tssp, (vm_offset_t)tssp +
|
pmap_pti_add_kva((vm_offset_t)tssp, (vm_offset_t)tssp +
|
||||||
ctob(IOPAGES + 1), false);
|
ctob(IOPAGES + 1), false);
|
||||||
iomap = (char *)&tssp[1];
|
iomap = (char *)&tssp[1];
|
||||||
@ -463,7 +463,7 @@ user_ldt_alloc(struct proc *p, int force)
|
|||||||
mtx_unlock(&dt_lock);
|
mtx_unlock(&dt_lock);
|
||||||
new_ldt = malloc(sizeof(struct proc_ldt), M_SUBPROC, M_WAITOK);
|
new_ldt = malloc(sizeof(struct proc_ldt), M_SUBPROC, M_WAITOK);
|
||||||
sz = max_ldt_segment * sizeof(struct user_segment_descriptor);
|
sz = max_ldt_segment * sizeof(struct user_segment_descriptor);
|
||||||
sva = kmem_malloc(kernel_arena, sz, M_WAITOK | M_ZERO);
|
sva = kmem_malloc(sz, M_WAITOK | M_ZERO);
|
||||||
new_ldt->ldt_base = (caddr_t)sva;
|
new_ldt->ldt_base = (caddr_t)sva;
|
||||||
pmap_pti_add_kva(sva, sva + sz, false);
|
pmap_pti_add_kva(sva, sva + sz, false);
|
||||||
new_ldt->ldt_refcnt = 1;
|
new_ldt->ldt_refcnt = 1;
|
||||||
|
@ -119,8 +119,7 @@ cpu_mp_start(void)
|
|||||||
|
|
||||||
/* Reserve memory for application processors */
|
/* Reserve memory for application processors */
|
||||||
for(i = 0; i < (mp_ncpus - 1); i++)
|
for(i = 0; i < (mp_ncpus - 1); i++)
|
||||||
dpcpu[i] = (void *)kmem_malloc(kernel_arena, DPCPU_SIZE,
|
dpcpu[i] = (void *)kmem_malloc(DPCPU_SIZE, M_WAITOK | M_ZERO);
|
||||||
M_WAITOK | M_ZERO);
|
|
||||||
|
|
||||||
dcache_wbinv_poc_all();
|
dcache_wbinv_poc_all();
|
||||||
|
|
||||||
|
@ -1782,8 +1782,7 @@ pmap_init(void)
|
|||||||
*/
|
*/
|
||||||
s = (vm_size_t)(pv_npg * sizeof(struct md_page));
|
s = (vm_size_t)(pv_npg * sizeof(struct md_page));
|
||||||
s = round_page(s);
|
s = round_page(s);
|
||||||
pv_table = (struct md_page *)kmem_malloc(kernel_arena, s,
|
pv_table = (struct md_page *)kmem_malloc(s, M_WAITOK | M_ZERO);
|
||||||
M_WAITOK | M_ZERO);
|
|
||||||
for (i = 0; i < pv_npg; i++)
|
for (i = 0; i < pv_npg; i++)
|
||||||
TAILQ_INIT(&pv_table[i].pv_list);
|
TAILQ_INIT(&pv_table[i].pv_list);
|
||||||
|
|
||||||
|
@ -484,8 +484,7 @@ start_cpu(u_int id, uint64_t target_cpu)
|
|||||||
pcpup = &__pcpu[cpuid];
|
pcpup = &__pcpu[cpuid];
|
||||||
pcpu_init(pcpup, cpuid, sizeof(struct pcpu));
|
pcpu_init(pcpup, cpuid, sizeof(struct pcpu));
|
||||||
|
|
||||||
dpcpu[cpuid - 1] = (void *)kmem_malloc(kernel_arena, DPCPU_SIZE,
|
dpcpu[cpuid - 1] = (void *)kmem_malloc(DPCPU_SIZE, M_WAITOK | M_ZERO);
|
||||||
M_WAITOK | M_ZERO);
|
|
||||||
dpcpu_init(dpcpu[cpuid - 1], cpuid);
|
dpcpu_init(dpcpu[cpuid - 1], cpuid);
|
||||||
|
|
||||||
printf("Starting CPU %u (%lx)\n", cpuid, target_cpu);
|
printf("Starting CPU %u (%lx)\n", cpuid, target_cpu);
|
||||||
|
@ -923,8 +923,7 @@ pmap_init(void)
|
|||||||
*/
|
*/
|
||||||
s = (vm_size_t)(pv_npg * sizeof(struct md_page));
|
s = (vm_size_t)(pv_npg * sizeof(struct md_page));
|
||||||
s = round_page(s);
|
s = round_page(s);
|
||||||
pv_table = (struct md_page *)kmem_malloc(kernel_arena, s,
|
pv_table = (struct md_page *)kmem_malloc(s, M_WAITOK | M_ZERO);
|
||||||
M_WAITOK | M_ZERO);
|
|
||||||
for (i = 0; i < pv_npg; i++)
|
for (i = 0; i < pv_npg; i++)
|
||||||
TAILQ_INIT(&pv_table[i].pv_list);
|
TAILQ_INIT(&pv_table[i].pv_list);
|
||||||
TAILQ_INIT(&pv_dummy.pv_list);
|
TAILQ_INIT(&pv_dummy.pv_list);
|
||||||
|
@ -165,7 +165,7 @@ linux_alloc_kmem(gfp_t flags, unsigned int order)
|
|||||||
vm_offset_t addr;
|
vm_offset_t addr;
|
||||||
|
|
||||||
if ((flags & GFP_DMA32) == 0) {
|
if ((flags & GFP_DMA32) == 0) {
|
||||||
addr = kmem_malloc(kmem_arena, size, flags & GFP_NATIVE_MASK);
|
addr = kmem_malloc(size, flags & GFP_NATIVE_MASK);
|
||||||
} else {
|
} else {
|
||||||
addr = kmem_alloc_contig(size, flags & GFP_NATIVE_MASK, 0,
|
addr = kmem_alloc_contig(size, flags & GFP_NATIVE_MASK, 0,
|
||||||
BUS_SPACE_MAXADDR_32BIT, PAGE_SIZE, 0, VM_MEMATTR_DEFAULT);
|
BUS_SPACE_MAXADDR_32BIT, PAGE_SIZE, 0, VM_MEMATTR_DEFAULT);
|
||||||
|
@ -283,8 +283,8 @@ hypercall_create(void *arg __unused)
|
|||||||
* the NX bit.
|
* the NX bit.
|
||||||
* - Assume kmem_malloc() returns properly aligned memory.
|
* - Assume kmem_malloc() returns properly aligned memory.
|
||||||
*/
|
*/
|
||||||
hypercall_context.hc_addr = (void *)kmem_malloc(kernel_arena, PAGE_SIZE,
|
hypercall_context.hc_addr = (void *)kmem_malloc(PAGE_SIZE, M_EXEC |
|
||||||
M_EXEC | M_WAITOK);
|
M_WAITOK);
|
||||||
hypercall_context.hc_paddr = vtophys(hypercall_context.hc_addr);
|
hypercall_context.hc_paddr = vtophys(hypercall_context.hc_addr);
|
||||||
|
|
||||||
/* Get the 'reserved' bits, which requires preservation. */
|
/* Get the 'reserved' bits, which requires preservation. */
|
||||||
|
@ -328,11 +328,9 @@ start_all_aps(void)
|
|||||||
apic_id = cpu_apic_ids[cpu];
|
apic_id = cpu_apic_ids[cpu];
|
||||||
|
|
||||||
/* allocate and set up a boot stack data page */
|
/* allocate and set up a boot stack data page */
|
||||||
bootstacks[cpu] =
|
bootstacks[cpu] = (char *)kmem_malloc(kstack_pages * PAGE_SIZE,
|
||||||
(char *)kmem_malloc(kernel_arena, kstack_pages * PAGE_SIZE,
|
|
||||||
M_WAITOK | M_ZERO);
|
|
||||||
dpcpu = (void *)kmem_malloc(kernel_arena, DPCPU_SIZE,
|
|
||||||
M_WAITOK | M_ZERO);
|
M_WAITOK | M_ZERO);
|
||||||
|
dpcpu = (void *)kmem_malloc(DPCPU_SIZE, M_WAITOK | M_ZERO);
|
||||||
/* setup a vector to our boot code */
|
/* setup a vector to our boot code */
|
||||||
*((volatile u_short *) WARMBOOT_OFF) = WARMBOOT_TARGET;
|
*((volatile u_short *) WARMBOOT_OFF) = WARMBOOT_TARGET;
|
||||||
*((volatile u_short *) WARMBOOT_SEG) = (boot_address >> 4);
|
*((volatile u_short *) WARMBOOT_SEG) = (boot_address >> 4);
|
||||||
|
@ -998,8 +998,7 @@ pmap_init(void)
|
|||||||
*/
|
*/
|
||||||
s = (vm_size_t)(pv_npg * sizeof(struct md_page));
|
s = (vm_size_t)(pv_npg * sizeof(struct md_page));
|
||||||
s = round_page(s);
|
s = round_page(s);
|
||||||
pv_table = (struct md_page *)kmem_malloc(kernel_arena, s,
|
pv_table = (struct md_page *)kmem_malloc(s, M_WAITOK | M_ZERO);
|
||||||
M_WAITOK | M_ZERO);
|
|
||||||
for (i = 0; i < pv_npg; i++)
|
for (i = 0; i < pv_npg; i++)
|
||||||
TAILQ_INIT(&pv_table[i].pv_list);
|
TAILQ_INIT(&pv_table[i].pv_list);
|
||||||
|
|
||||||
|
@ -185,7 +185,7 @@ start_ap(int cpuid)
|
|||||||
int cpus, ms;
|
int cpus, ms;
|
||||||
|
|
||||||
cpus = mp_naps;
|
cpus = mp_naps;
|
||||||
dpcpu = (void *)kmem_malloc(kernel_arena, DPCPU_SIZE, M_WAITOK | M_ZERO);
|
dpcpu = (void *)kmem_malloc(DPCPU_SIZE, M_WAITOK | M_ZERO);
|
||||||
|
|
||||||
mips_sync();
|
mips_sync();
|
||||||
|
|
||||||
|
@ -168,8 +168,8 @@ cpu_mp_start(void)
|
|||||||
void *dpcpu;
|
void *dpcpu;
|
||||||
|
|
||||||
pc = &__pcpu[cpu.cr_cpuid];
|
pc = &__pcpu[cpu.cr_cpuid];
|
||||||
dpcpu = (void *)kmem_malloc(kernel_arena, DPCPU_SIZE,
|
dpcpu = (void *)kmem_malloc(DPCPU_SIZE, M_WAITOK |
|
||||||
M_WAITOK | M_ZERO);
|
M_ZERO);
|
||||||
pcpu_init(pc, cpu.cr_cpuid, sizeof(*pc));
|
pcpu_init(pc, cpu.cr_cpuid, sizeof(*pc));
|
||||||
dpcpu_init(dpcpu, cpu.cr_cpuid);
|
dpcpu_init(dpcpu, cpu.cr_cpuid);
|
||||||
} else {
|
} else {
|
||||||
|
@ -391,8 +391,7 @@ cpu_init_fdt(u_int id, phandle_t node, u_int addr_size, pcell_t *reg)
|
|||||||
|
|
||||||
pcpu_init(pcpup, id, sizeof(struct pcpu));
|
pcpu_init(pcpup, id, sizeof(struct pcpu));
|
||||||
|
|
||||||
dpcpu[id - 1] = (void *)kmem_malloc(kernel_arena, DPCPU_SIZE,
|
dpcpu[id - 1] = (void *)kmem_malloc(DPCPU_SIZE, M_WAITOK | M_ZERO);
|
||||||
M_WAITOK | M_ZERO);
|
|
||||||
dpcpu_init(dpcpu[id - 1], id);
|
dpcpu_init(dpcpu[id - 1], id);
|
||||||
|
|
||||||
printf("Starting CPU %u (%lx)\n", id, target_cpu);
|
printf("Starting CPU %u (%lx)\n", id, target_cpu);
|
||||||
|
@ -342,12 +342,10 @@ ap_start(phandle_t node, u_int mid, u_int cpu_impl)
|
|||||||
cpuid_to_mid[cpuid] = mid;
|
cpuid_to_mid[cpuid] = mid;
|
||||||
cpu_identify(csa->csa_ver, clock, cpuid);
|
cpu_identify(csa->csa_ver, clock, cpuid);
|
||||||
|
|
||||||
va = kmem_malloc(kernel_arena, PCPU_PAGES * PAGE_SIZE,
|
va = kmem_malloc(PCPU_PAGES * PAGE_SIZE, M_WAITOK | M_ZERO);
|
||||||
M_WAITOK | M_ZERO);
|
|
||||||
pc = (struct pcpu *)(va + (PCPU_PAGES * PAGE_SIZE)) - 1;
|
pc = (struct pcpu *)(va + (PCPU_PAGES * PAGE_SIZE)) - 1;
|
||||||
pcpu_init(pc, cpuid, sizeof(*pc));
|
pcpu_init(pc, cpuid, sizeof(*pc));
|
||||||
dpcpu_init((void *)kmem_malloc(kernel_arena, DPCPU_SIZE,
|
dpcpu_init((void *)kmem_malloc(DPCPU_SIZE, M_WAITOK | M_ZERO), cpuid);
|
||||||
M_WAITOK | M_ZERO), cpuid);
|
|
||||||
pc->pc_addr = va;
|
pc->pc_addr = va;
|
||||||
pc->pc_clock = clock;
|
pc->pc_clock = clock;
|
||||||
pc->pc_impl = cpu_impl;
|
pc->pc_impl = cpu_impl;
|
||||||
|
@ -3687,7 +3687,7 @@ uma_large_malloc_domain(vm_size_t size, int domain, int wait)
|
|||||||
if (slab == NULL)
|
if (slab == NULL)
|
||||||
return (NULL);
|
return (NULL);
|
||||||
if (domain == UMA_ANYDOMAIN)
|
if (domain == UMA_ANYDOMAIN)
|
||||||
addr = kmem_malloc(NULL, size, wait);
|
addr = kmem_malloc(size, wait);
|
||||||
else
|
else
|
||||||
addr = kmem_malloc_domain(domain, size, wait);
|
addr = kmem_malloc_domain(domain, size, wait);
|
||||||
if (addr != 0) {
|
if (addr != 0) {
|
||||||
|
@ -64,7 +64,7 @@ vm_offset_t kmem_alloc_contig(vm_size_t size, int flags,
|
|||||||
vm_offset_t kmem_alloc_contig_domain(int domain, vm_size_t size, int flags,
|
vm_offset_t kmem_alloc_contig_domain(int domain, vm_size_t size, int flags,
|
||||||
vm_paddr_t low, vm_paddr_t high, u_long alignment, vm_paddr_t boundary,
|
vm_paddr_t low, vm_paddr_t high, u_long alignment, vm_paddr_t boundary,
|
||||||
vm_memattr_t memattr);
|
vm_memattr_t memattr);
|
||||||
vm_offset_t kmem_malloc(struct vmem *, vm_size_t size, int flags);
|
vm_offset_t kmem_malloc(vm_size_t size, int flags);
|
||||||
vm_offset_t kmem_malloc_domain(int domain, vm_size_t size, int flags);
|
vm_offset_t kmem_malloc_domain(int domain, vm_size_t size, int flags);
|
||||||
void kmem_free(struct vmem *, vm_offset_t, vm_size_t);
|
void kmem_free(struct vmem *, vm_offset_t, vm_size_t);
|
||||||
|
|
||||||
|
@ -278,8 +278,7 @@ vm_ksubmap_init(struct kva_md_info *kmi)
|
|||||||
(vm_paddr_t)1 << 32, ~(vm_paddr_t)0, VM_MEMATTR_DEFAULT);
|
(vm_paddr_t)1 << 32, ~(vm_paddr_t)0, VM_MEMATTR_DEFAULT);
|
||||||
if (firstaddr == 0)
|
if (firstaddr == 0)
|
||||||
#endif
|
#endif
|
||||||
firstaddr = kmem_malloc(kernel_arena, size,
|
firstaddr = kmem_malloc(size, M_ZERO | M_WAITOK);
|
||||||
M_ZERO | M_WAITOK);
|
|
||||||
if (firstaddr == 0)
|
if (firstaddr == 0)
|
||||||
panic("startup: no room for tables");
|
panic("startup: no room for tables");
|
||||||
goto again;
|
goto again;
|
||||||
|
@ -392,7 +392,7 @@ kmem_malloc_domain(int domain, vm_size_t size, int flags)
|
|||||||
}
|
}
|
||||||
|
|
||||||
vm_offset_t
|
vm_offset_t
|
||||||
kmem_malloc(struct vmem *vmem __unused, vm_size_t size, int flags)
|
kmem_malloc(vm_size_t size, int flags)
|
||||||
{
|
{
|
||||||
struct vm_domainset_iter di;
|
struct vm_domainset_iter di;
|
||||||
vm_offset_t addr;
|
vm_offset_t addr;
|
||||||
|
@ -350,18 +350,12 @@ start_xen_ap(int cpu)
|
|||||||
const size_t stacksize = kstack_pages * PAGE_SIZE;
|
const size_t stacksize = kstack_pages * PAGE_SIZE;
|
||||||
|
|
||||||
/* allocate and set up an idle stack data page */
|
/* allocate and set up an idle stack data page */
|
||||||
bootstacks[cpu] =
|
bootstacks[cpu] = (void *)kmem_malloc(stacksize, M_WAITOK | M_ZERO);
|
||||||
(void *)kmem_malloc(kernel_arena, stacksize, M_WAITOK | M_ZERO);
|
doublefault_stack = (char *)kmem_malloc(PAGE_SIZE, M_WAITOK | M_ZERO);
|
||||||
doublefault_stack =
|
mce_stack = (char *)kmem_malloc(PAGE_SIZE, M_WAITOK | M_ZERO);
|
||||||
(char *)kmem_malloc(kernel_arena, PAGE_SIZE, M_WAITOK | M_ZERO);
|
nmi_stack = (char *)kmem_malloc(PAGE_SIZE, M_WAITOK | M_ZERO);
|
||||||
mce_stack =
|
dbg_stack = (void *)kmem_malloc(PAGE_SIZE, M_WAITOK | M_ZERO);
|
||||||
(char *)kmem_malloc(kernel_arena, PAGE_SIZE, M_WAITOK | M_ZERO);
|
dpcpu = (void *)kmem_malloc(DPCPU_SIZE, M_WAITOK | M_ZERO);
|
||||||
nmi_stack =
|
|
||||||
(char *)kmem_malloc(kernel_arena, PAGE_SIZE, M_WAITOK | M_ZERO);
|
|
||||||
dbg_stack =
|
|
||||||
(void *)kmem_malloc(kernel_arena, PAGE_SIZE, M_WAITOK | M_ZERO);
|
|
||||||
dpcpu =
|
|
||||||
(void *)kmem_malloc(kernel_arena, DPCPU_SIZE, M_WAITOK | M_ZERO);
|
|
||||||
|
|
||||||
bootSTK = (char *)bootstacks[cpu] + kstack_pages * PAGE_SIZE - 8;
|
bootSTK = (char *)bootstacks[cpu] + kstack_pages * PAGE_SIZE - 8;
|
||||||
bootAP = cpu;
|
bootAP = cpu;
|
||||||
|
Loading…
Reference in New Issue
Block a user