o) Add a VM find-space option, VMFS_TLB_ALIGNED_SPACE, which searches the

address space for an address as aligned by the new pmap_align_tlb()
   function, which is for constraints imposed by the TLB. [1]
o) Add a kmem_alloc_nofault_space() function, which acts like
   kmem_alloc_nofault() but allows the caller to specify which find-space
   option to use. [1]
o) Use kmem_alloc_nofault_space() with VMFS_TLB_ALIGNED_SPACE to allocate the
   kernel stack address on MIPS. [1]
o) Make pmap_align_tlb() on MIPS align addresses so that they do not start on
   an odd boundary within the TLB, so that they are suitable for insertion as
   wired entries and do not have to share a TLB entry with another mapping,
   assuming they are appropriately-sized.
o) Eliminate md_realstack now that the kstack will be appropriately-aligned on
   MIPS.
o) Increase the number of guard pages to 2 so that we retain the proper
   alignment of the kstack address.

Reviewed by:	[1] alc
X-MFC-after:	Making sure alc has not come up with a better interface.
This commit is contained in:
Juli Mallett 2010-04-18 22:32:07 +00:00
parent 7ea710b3b1
commit ca596a25f0
Notes: svn2git 2020-12-20 02:59:44 +00:00
svn path=/head/; revision=206819
14 changed files with 88 additions and 34 deletions

View File

@ -113,12 +113,9 @@
/*
* The kernel stack needs to be aligned on a (PAGE_SIZE * 2) boundary.
*
* Although we allocate 3 pages for the kernel stack we end up using
* only the 2 pages that are aligned on a (PAGE_SIZE * 2) boundary.
*/
#define KSTACK_PAGES 3 /* kernel stack*/
#define KSTACK_GUARD_PAGES 1 /* pages of kstack guard; 0 disables */
#define KSTACK_PAGES 2 /* kernel stack*/
#define KSTACK_GUARD_PAGES 2 /* pages of kstack guard; 0 disables */
#define UPAGES 2

View File

@ -44,7 +44,7 @@
*/
struct mdthread {
int md_flags; /* machine-dependent flags */
int md_upte[KSTACK_PAGES - 1]; /* ptes for mapping u pcb */
int md_upte[KSTACK_PAGES]; /* ptes for mapping u pcb */
int md_ss_addr; /* single step address for ptrace */
int md_ss_instr; /* single step instruction for ptrace */
register_t md_saved_intr;
@ -53,7 +53,6 @@ struct mdthread {
int md_pc_ctrl; /* performance counter control */
int md_pc_count; /* performance counter */
int md_pc_spill; /* performance counter spill */
vm_offset_t md_realstack;
void *md_tls;
};

View File

@ -928,7 +928,7 @@ tlb_insert_random:
*/
GET_CPU_PCPU(k1)
lw k0, PC_CURTHREAD(k1)
lw k0, TD_REALKSTACK(k0)
lw k0, TD_KSTACK(k0)
sltu k0, k0, sp
bnez k0, _C_LABEL(MipsKernGenException)
nop
@ -975,7 +975,7 @@ tlb_insert_random:
*/
GET_CPU_PCPU(k1)
lw k0, PC_CURTHREAD(k1)
sw zero, TD_REALKSTACK(k0)
sw zero, TD_KSTACK(k0)
move a1, a0
PANIC("kernel stack overflow - trapframe at %p")

View File

@ -65,7 +65,7 @@ __FBSDID("$FreeBSD$");
ASSYM(TD_PCB, offsetof(struct thread, td_pcb));
ASSYM(TD_UPTE, offsetof(struct thread, td_md.md_upte));
ASSYM(TD_REALKSTACK, offsetof(struct thread, td_md.md_realstack));
ASSYM(TD_KSTACK, offsetof(struct thread, td_kstack));
ASSYM(TD_FLAGS, offsetof(struct thread, td_flags));
ASSYM(TD_LOCK, offsetof(struct thread, td_lock));
ASSYM(TD_FRAME, offsetof(struct thread, td_frame));

View File

@ -298,14 +298,13 @@ mips_proc0_init(void)
(long)kstack0));
thread0.td_kstack = kstack0;
thread0.td_kstack_pages = KSTACK_PAGES;
thread0.td_md.md_realstack = roundup2(thread0.td_kstack, PAGE_SIZE * 2);
/*
* Do not use cpu_thread_alloc to initialize these fields
* thread0 is the only thread that has kstack located in KSEG0
* while cpu_thread_alloc handles kstack allocated in KSEG2.
*/
thread0.td_pcb = (struct pcb *)(thread0.td_md.md_realstack +
(thread0.td_kstack_pages - 1) * PAGE_SIZE) - 1;
thread0.td_pcb = (struct pcb *)(thread0.td_kstack +
thread0.td_kstack_pages * PAGE_SIZE) - 1;
thread0.td_frame = &thread0.td_pcb->pcb_regs;
/* Steal memory for the dynamic per-cpu area. */

View File

@ -2813,6 +2813,21 @@ pmap_align_superpage(vm_object_t object, vm_ooffset_t offset,
*addr = ((*addr + SEGOFSET) & ~SEGOFSET) + superpage_offset;
}
/*
* Increase the starting virtual address of the given mapping so
* that it is aligned to not be the second page in a TLB entry.
* This routine assumes that the length is appropriately-sized so
* that the allocation does not share a TLB entry at all if required.
*/
void
pmap_align_tlb(vm_offset_t *addr)
{
if ((*addr & PAGE_SIZE) == 0)
return;
*addr += PAGE_SIZE;
return;
}
int pmap_pid_dump(int pid);
int

View File

@ -339,7 +339,7 @@ blocked_loop:
sw a1, PC_CURTHREAD(a3)
lw a2, TD_PCB(a1)
sw a2, PC_CURPCB(a3)
lw v0, TD_REALKSTACK(a1)
lw v0, TD_KSTACK(a1)
li s0, MIPS_KSEG2_START # If Uarea addr is below kseg2,
bltu v0, s0, sw2 # no need to insert in TLB.
lw a1, TD_UPTE+0(s7) # t0 = first u. pte

View File

@ -217,13 +217,9 @@ cpu_thread_swapin(struct thread *td)
* part of the thread struct so cpu_switch() can quickly map in
* the pcb struct and kernel stack.
*/
if (!(pte = pmap_segmap(kernel_pmap, td->td_md.md_realstack)))
panic("cpu_thread_swapin: invalid segmap");
pte += ((vm_offset_t)td->td_md.md_realstack >> PAGE_SHIFT) & (NPTEPG - 1);
for (i = 0; i < KSTACK_PAGES - 1; i++) {
for (i = 0; i < KSTACK_PAGES; i++) {
pte = pmap_pte(kernel_pmap, td->td_kstack + i * PAGE_SIZE);
td->td_md.md_upte[i] = *pte & ~(PTE_RO|PTE_WIRED);
pte++;
}
}
@ -238,22 +234,14 @@ cpu_thread_alloc(struct thread *td)
pt_entry_t *pte;
int i;
if (td->td_kstack & (1 << PAGE_SHIFT))
td->td_md.md_realstack = td->td_kstack + PAGE_SIZE;
else
td->td_md.md_realstack = td->td_kstack;
td->td_pcb = (struct pcb *)(td->td_md.md_realstack +
(td->td_kstack_pages - 1) * PAGE_SIZE) - 1;
KASSERT((td->td_kstack & (1 << PAGE_SHIFT)) == 0, ("kernel stack must be aligned."));
td->td_pcb = (struct pcb *)(td->td_kstack +
td->td_kstack_pages * PAGE_SIZE) - 1;
td->td_frame = &td->td_pcb->pcb_regs;
if (!(pte = pmap_segmap(kernel_pmap, td->td_md.md_realstack)))
panic("cpu_thread_alloc: invalid segmap");
pte += ((vm_offset_t)td->td_md.md_realstack >> PAGE_SHIFT) & (NPTEPG - 1);
for (i = 0; i < KSTACK_PAGES - 1; i++) {
for (i = 0; i < KSTACK_PAGES; i++) {
pte = pmap_pte(kernel_pmap, td->td_kstack + i * PAGE_SIZE);
td->td_md.md_upte[i] = *pte & ~(PTE_RO|PTE_WIRED);
pte++;
}
}

View File

@ -98,6 +98,9 @@ extern vm_offset_t kernel_vm_end;
void pmap_align_superpage(vm_object_t, vm_ooffset_t, vm_offset_t *,
vm_size_t);
#if defined(__mips__)
void pmap_align_tlb(vm_offset_t *);
#endif
void pmap_change_wiring(pmap_t, vm_offset_t, boolean_t);
void pmap_clear_modify(vm_page_t m);
void pmap_clear_reference(vm_page_t m);

View File

@ -47,6 +47,7 @@ vm_offset_t kmem_alloc_contig(vm_map_t map, vm_size_t size, int flags,
vm_paddr_t low, vm_paddr_t high, unsigned long alignment,
unsigned long boundary, vm_memattr_t memattr);
vm_offset_t kmem_alloc_nofault(vm_map_t, vm_size_t);
vm_offset_t kmem_alloc_nofault_space(vm_map_t, vm_size_t, int);
vm_offset_t kmem_alloc_wait(vm_map_t, vm_size_t);
void kmem_free(vm_map_t, vm_offset_t, vm_size_t);
void kmem_free_wakeup(vm_map_t, vm_offset_t, vm_size_t);

View File

@ -373,8 +373,17 @@ vm_thread_new(struct thread *td, int pages)
/*
* Get a kernel virtual address for this thread's kstack.
*/
#if defined(__mips__)
/*
* We need to align the kstack's mapped address to fit within
* a single TLB entry.
*/
ks = kmem_alloc_nofault_space(kernel_map,
(pages + KSTACK_GUARD_PAGES) * PAGE_SIZE, VMFS_TLB_ALIGNED_SPACE);
#else
ks = kmem_alloc_nofault(kernel_map,
(pages + KSTACK_GUARD_PAGES) * PAGE_SIZE);
#endif
if (ks == 0) {
printf("vm_thread_new: kstack allocation failed\n");
vm_object_deallocate(ksobj);

View File

@ -118,6 +118,35 @@ kmem_alloc_nofault(map, size)
return (addr);
}
/*
* kmem_alloc_nofault_space:
*
* Allocate a virtual address range with no underlying object and
* no initial mapping to physical memory within the specified
* address space. Any mapping from this range to physical memory
* must be explicitly created prior to its use, typically with
* pmap_qenter(). Any attempt to create a mapping on demand
* through vm_fault() will result in a panic.
*/
vm_offset_t
kmem_alloc_nofault_space(map, size, find_space)
vm_map_t map;
vm_size_t size;
int find_space;
{
vm_offset_t addr;
int result;
size = round_page(size);
addr = vm_map_min(map);
result = vm_map_find(map, NULL, 0, &addr, size, find_space,
VM_PROT_ALL, VM_PROT_ALL, MAP_NOFAULT);
if (result != KERN_SUCCESS) {
return (0);
}
return (addr);
}
/*
* Allocate wired-down memory in the kernel's address map
* or a submap.

View File

@ -1394,9 +1394,20 @@ vm_map_find(vm_map_t map, vm_object_t object, vm_ooffset_t offset,
vm_map_unlock(map);
return (KERN_NO_SPACE);
}
if (find_space == VMFS_ALIGNED_SPACE)
switch (find_space) {
case VMFS_ALIGNED_SPACE:
pmap_align_superpage(object, offset, addr,
length);
break;
#ifdef VMFS_TLB_ALIGNED_SPACE
case VMFS_TLB_ALIGNED_SPACE:
pmap_align_tlb(addr);
break;
#endif
default:
break;
}
start = *addr;
}
result = vm_map_insert(map, object, offset, start, start +

View File

@ -326,6 +326,9 @@ long vmspace_wired_count(struct vmspace *vmspace);
#define VMFS_NO_SPACE 0 /* don't find; use the given range */
#define VMFS_ANY_SPACE 1 /* find a range with any alignment */
#define VMFS_ALIGNED_SPACE 2 /* find a superpage-aligned range */
#if defined(__mips__)
#define VMFS_TLB_ALIGNED_SPACE 3 /* find a TLB entry aligned range */
#endif
/*
* vm_map_wire and vm_map_unwire option flags