d22883d715
e4b8deb222
removed the last in-tree uses of PCPU_INC(). Its
potential benefit is also practically nonexistent. Non-x86
platforms already implement it as PCPU_ADD(..., 1), and according
to [0] there are no recent x86 processors for which the 'inc'
instruction provides a performance benefit over the equivalent
memory-operand form of the 'add' instruction. The only remaining
benefit of 'inc' is smaller instruction size, which in this case
is inconsequential given the limited number of per-CPU data consumers.
[0]: https://www.agner.org/optimize/instruction_tables.pdf
Reviewed by: kib
Differential Revision: https://reviews.freebsd.org/D29308
289 lines
9.0 KiB
C
289 lines
9.0 KiB
C
/*-
|
|
* SPDX-License-Identifier: BSD-2-Clause-FreeBSD
|
|
*
|
|
* Copyright (c) Peter Wemm <peter@netplex.com.au>
|
|
* All rights reserved.
|
|
*
|
|
* Redistribution and use in source and binary forms, with or without
|
|
* modification, are permitted provided that the following conditions
|
|
* are met:
|
|
* 1. Redistributions of source code must retain the above copyright
|
|
* notice, this list of conditions and the following disclaimer.
|
|
* 2. Redistributions in binary form must reproduce the above copyright
|
|
* notice, this list of conditions and the following disclaimer in the
|
|
* documentation and/or other materials provided with the distribution.
|
|
*
|
|
* THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND
|
|
* ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
|
|
* IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
|
|
* ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE
|
|
* FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
|
|
* DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
|
|
* OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
|
|
* HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
|
|
* LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
|
|
* OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
|
|
* SUCH DAMAGE.
|
|
*
|
|
* $FreeBSD$
|
|
*/
|
|
|
|
#ifndef _MACHINE_PCPU_H_
|
|
#define _MACHINE_PCPU_H_
|
|
|
|
#ifndef _SYS_CDEFS_H_
|
|
#error "sys/cdefs.h is a prerequisite for this file"
|
|
#endif
|
|
|
|
#include <machine/segments.h>
|
|
#include <machine/tss.h>
|
|
|
|
#define PC_PTI_STACK_SZ 16
|
|
|
|
struct monitorbuf {
|
|
int idle_state; /* Used by cpu_idle_mwait. */
|
|
int stop_state; /* Used by cpustop_handler. */
|
|
char padding[128 - (2 * sizeof(int))];
|
|
};
|
|
_Static_assert(sizeof(struct monitorbuf) == 128, "2x cache line");
|
|
|
|
/*
|
|
* The SMP parts are setup in pmap.c and locore.s for the BSP, and
|
|
* mp_machdep.c sets up the data for the AP's to "see" when they awake.
|
|
* The reason for doing it via a struct is so that an array of pointers
|
|
* to each CPU's data can be set up for things like "check curproc on all
|
|
* other processors"
|
|
*/
|
|
#define PCPU_MD_FIELDS \
|
|
struct monitorbuf pc_monitorbuf __aligned(128); /* cache line */\
|
|
struct pcpu *pc_prvspace; /* Self-reference */ \
|
|
struct pmap *pc_curpmap; \
|
|
struct amd64tss *pc_tssp; /* TSS segment active on CPU */ \
|
|
void *pc_pad0; \
|
|
uint64_t pc_kcr3; \
|
|
uint64_t pc_ucr3; \
|
|
uint64_t pc_saved_ucr3; \
|
|
register_t pc_rsp0; \
|
|
register_t pc_scratch_rsp; /* User %rsp in syscall */ \
|
|
register_t pc_scratch_rax; \
|
|
u_int pc_apic_id; \
|
|
u_int pc_acpi_id; /* ACPI CPU id */ \
|
|
/* Pointer to the CPU %fs descriptor */ \
|
|
struct user_segment_descriptor *pc_fs32p; \
|
|
/* Pointer to the CPU %gs descriptor */ \
|
|
struct user_segment_descriptor *pc_gs32p; \
|
|
/* Pointer to the CPU LDT descriptor */ \
|
|
struct system_segment_descriptor *pc_ldt; \
|
|
/* Pointer to the CPU TSS descriptor */ \
|
|
struct system_segment_descriptor *pc_tss; \
|
|
u_int pc_cmci_mask; /* MCx banks for CMCI */ \
|
|
uint64_t pc_dbreg[16]; /* ddb debugging regs */ \
|
|
uint64_t pc_pti_stack[PC_PTI_STACK_SZ]; \
|
|
register_t pc_pti_rsp0; \
|
|
int pc_dbreg_cmd; /* ddb debugging reg cmd */ \
|
|
u_int pc_vcpu_id; /* Xen vCPU ID */ \
|
|
uint32_t pc_pcid_next; \
|
|
uint32_t pc_pcid_gen; \
|
|
uint32_t pc_unused; \
|
|
uint32_t pc_ibpb_set; \
|
|
void *pc_mds_buf; \
|
|
void *pc_mds_buf64; \
|
|
uint32_t pc_pad[4]; \
|
|
uint8_t pc_mds_tmp[64]; \
|
|
u_int pc_ipi_bitmap; \
|
|
struct amd64tss pc_common_tss; \
|
|
struct user_segment_descriptor pc_gdt[NGDT]; \
|
|
void *pc_smp_tlb_pmap; \
|
|
uint64_t pc_smp_tlb_addr1; \
|
|
uint64_t pc_smp_tlb_addr2; \
|
|
uint32_t pc_smp_tlb_gen; \
|
|
u_int pc_smp_tlb_op; \
|
|
uint64_t pc_ucr3_load_mask; \
|
|
char __pad[2916] /* pad to UMA_PCPU_ALLOC_SIZE */
|
|
|
|
#define PC_DBREG_CMD_NONE 0
|
|
#define PC_DBREG_CMD_LOAD 1
|
|
|
|
#ifdef _KERNEL
|
|
|
|
#define MONITOR_STOPSTATE_RUNNING 0
|
|
#define MONITOR_STOPSTATE_STOPPED 1
|
|
|
|
#if defined(__GNUCLIKE_ASM) && defined(__GNUCLIKE___TYPEOF)
|
|
|
|
/*
|
|
* Evaluates to the byte offset of the per-cpu variable name.
|
|
*/
|
|
#define __pcpu_offset(name) \
|
|
__offsetof(struct pcpu, name)
|
|
|
|
/*
|
|
* Evaluates to the type of the per-cpu variable name.
|
|
*/
|
|
#define __pcpu_type(name) \
|
|
__typeof(((struct pcpu *)0)->name)
|
|
|
|
/*
|
|
* Evaluates to the address of the per-cpu variable name.
|
|
*/
|
|
#define __PCPU_PTR(name) __extension__ ({ \
|
|
__pcpu_type(name) *__p; \
|
|
\
|
|
__asm __volatile("movq %%gs:%1,%0; addq %2,%0" \
|
|
: "=r" (__p) \
|
|
: "m" (*(struct pcpu *)(__pcpu_offset(pc_prvspace))), \
|
|
"i" (__pcpu_offset(name))); \
|
|
\
|
|
__p; \
|
|
})
|
|
|
|
/*
|
|
* Evaluates to the value of the per-cpu variable name.
|
|
*/
|
|
#define __PCPU_GET(name) __extension__ ({ \
|
|
__pcpu_type(name) __res; \
|
|
struct __s { \
|
|
u_char __b[MIN(sizeof(__pcpu_type(name)), 8)]; \
|
|
} __s; \
|
|
\
|
|
if (sizeof(__res) == 1 || sizeof(__res) == 2 || \
|
|
sizeof(__res) == 4 || sizeof(__res) == 8) { \
|
|
__asm __volatile("mov %%gs:%1,%0" \
|
|
: "=r" (__s) \
|
|
: "m" (*(struct __s *)(__pcpu_offset(name)))); \
|
|
*(struct __s *)(void *)&__res = __s; \
|
|
} else { \
|
|
__res = *__PCPU_PTR(name); \
|
|
} \
|
|
__res; \
|
|
})
|
|
|
|
/*
|
|
* Adds the value to the per-cpu counter name. The implementation
|
|
* must be atomic with respect to interrupts.
|
|
*/
|
|
#define __PCPU_ADD(name, val) do { \
|
|
__pcpu_type(name) __val; \
|
|
struct __s { \
|
|
u_char __b[MIN(sizeof(__pcpu_type(name)), 8)]; \
|
|
} __s; \
|
|
\
|
|
__val = (val); \
|
|
if (sizeof(__val) == 1 || sizeof(__val) == 2 || \
|
|
sizeof(__val) == 4 || sizeof(__val) == 8) { \
|
|
__s = *(struct __s *)(void *)&__val; \
|
|
__asm __volatile("add %1,%%gs:%0" \
|
|
: "=m" (*(struct __s *)(__pcpu_offset(name))) \
|
|
: "r" (__s)); \
|
|
} else \
|
|
*__PCPU_PTR(name) += __val; \
|
|
} while (0)
|
|
|
|
/*
|
|
* Sets the value of the per-cpu variable name to value val.
|
|
*/
|
|
#define __PCPU_SET(name, val) { \
|
|
__pcpu_type(name) __val; \
|
|
struct __s { \
|
|
u_char __b[MIN(sizeof(__pcpu_type(name)), 8)]; \
|
|
} __s; \
|
|
\
|
|
__val = (val); \
|
|
if (sizeof(__val) == 1 || sizeof(__val) == 2 || \
|
|
sizeof(__val) == 4 || sizeof(__val) == 8) { \
|
|
__s = *(struct __s *)(void *)&__val; \
|
|
__asm __volatile("mov %1,%%gs:%0" \
|
|
: "=m" (*(struct __s *)(__pcpu_offset(name))) \
|
|
: "r" (__s)); \
|
|
} else { \
|
|
*__PCPU_PTR(name) = __val; \
|
|
} \
|
|
}
|
|
|
|
#define get_pcpu() __extension__ ({ \
|
|
struct pcpu *__pc; \
|
|
\
|
|
__asm __volatile("movq %%gs:%1,%0" \
|
|
: "=r" (__pc) \
|
|
: "m" (*(struct pcpu *)(__pcpu_offset(pc_prvspace)))); \
|
|
__pc; \
|
|
})
|
|
|
|
#define PCPU_GET(member) __PCPU_GET(pc_ ## member)
|
|
#define PCPU_ADD(member, val) __PCPU_ADD(pc_ ## member, val)
|
|
#define PCPU_PTR(member) __PCPU_PTR(pc_ ## member)
|
|
#define PCPU_SET(member, val) __PCPU_SET(pc_ ## member, val)
|
|
|
|
#define IS_BSP() (PCPU_GET(cpuid) == 0)
|
|
|
|
#define zpcpu_offset_cpu(cpu) ((uintptr_t)&__pcpu[0] + UMA_PCPU_ALLOC_SIZE * cpu)
|
|
#define zpcpu_base_to_offset(base) (void *)((uintptr_t)(base) - (uintptr_t)&__pcpu[0])
|
|
#define zpcpu_offset_to_base(base) (void *)((uintptr_t)(base) + (uintptr_t)&__pcpu[0])
|
|
|
|
#define zpcpu_sub_protected(base, n) do { \
|
|
ZPCPU_ASSERT_PROTECTED(); \
|
|
zpcpu_sub(base, n); \
|
|
} while (0)
|
|
|
|
#define zpcpu_set_protected(base, n) do { \
|
|
__typeof(*base) __n = (n); \
|
|
ZPCPU_ASSERT_PROTECTED(); \
|
|
switch (sizeof(*base)) { \
|
|
case 4: \
|
|
__asm __volatile("movl\t%1,%%gs:(%0)" \
|
|
: : "r" (base), "ri" (__n) : "memory", "cc"); \
|
|
break; \
|
|
case 8: \
|
|
__asm __volatile("movq\t%1,%%gs:(%0)" \
|
|
: : "r" (base), "ri" (__n) : "memory", "cc"); \
|
|
break; \
|
|
default: \
|
|
*zpcpu_get(base) = __n; \
|
|
} \
|
|
} while (0);
|
|
|
|
#define zpcpu_add(base, n) do { \
|
|
__typeof(*base) __n = (n); \
|
|
CTASSERT(sizeof(*base) == 4 || sizeof(*base) == 8); \
|
|
switch (sizeof(*base)) { \
|
|
case 4: \
|
|
__asm __volatile("addl\t%1,%%gs:(%0)" \
|
|
: : "r" (base), "ri" (__n) : "memory", "cc"); \
|
|
break; \
|
|
case 8: \
|
|
__asm __volatile("addq\t%1,%%gs:(%0)" \
|
|
: : "r" (base), "ri" (__n) : "memory", "cc"); \
|
|
break; \
|
|
} \
|
|
} while (0)
|
|
|
|
#define zpcpu_add_protected(base, n) do { \
|
|
ZPCPU_ASSERT_PROTECTED(); \
|
|
zpcpu_add(base, n); \
|
|
} while (0)
|
|
|
|
#define zpcpu_sub(base, n) do { \
|
|
__typeof(*base) __n = (n); \
|
|
CTASSERT(sizeof(*base) == 4 || sizeof(*base) == 8); \
|
|
switch (sizeof(*base)) { \
|
|
case 4: \
|
|
__asm __volatile("subl\t%1,%%gs:(%0)" \
|
|
: : "r" (base), "ri" (__n) : "memory", "cc"); \
|
|
break; \
|
|
case 8: \
|
|
__asm __volatile("subq\t%1,%%gs:(%0)" \
|
|
: : "r" (base), "ri" (__n) : "memory", "cc"); \
|
|
break; \
|
|
} \
|
|
} while (0);
|
|
|
|
#else /* !__GNUCLIKE_ASM || !__GNUCLIKE___TYPEOF */
|
|
|
|
#error "this file needs to be ported to your compiler"
|
|
|
|
#endif /* __GNUCLIKE_ASM && __GNUCLIKE___TYPEOF */
|
|
|
|
#endif /* _KERNEL */
|
|
|
|
#endif /* !_MACHINE_PCPU_H_ */
|