Reviewed and applied spl-01-rm-gpl-symbol-set_cpus_allowed.patch
from Ricardo which removes a dependency on the GPL-only symbol set_cpus_allowed(). Using this symbol is simpler but in the name of portability we are adopting a spinlock based solution here to remove this dependency. git-svn-id: https://outreach.scidac.gov/svn/spl/trunk@160 7e1ea52c-4ff2-0310-8f11-9dd32ca42a1c
This commit is contained in:
parent
d50bd9e221
commit
f6c81c5ea7
@ -157,6 +157,15 @@ union trace_data_union {
|
||||
unsigned short tcd_type;
|
||||
/* The factors to share debug memory. */
|
||||
unsigned short tcd_pages_factor;
|
||||
|
||||
/*
|
||||
* This spinlock is needed to workaround the problem of
|
||||
* set_cpus_allowed() being GPL-only. Since we cannot
|
||||
* schedule a thread on a specific CPU when dumping the
|
||||
* pages, we must use the spinlock for mutual exclusion.
|
||||
*/
|
||||
spinlock_t tcd_lock;
|
||||
unsigned long tcd_lock_flags;
|
||||
} tcd;
|
||||
char __pad[L1_CACHE_ALIGN(sizeof(struct trace_cpu_data))];
|
||||
};
|
||||
@ -168,9 +177,9 @@ extern union trace_data_union (*trace_data[TCD_TYPE_MAX])[NR_CPUS];
|
||||
for (j = 0, ((tcd) = &(*trace_data[i])[j].tcd); \
|
||||
j < num_possible_cpus(); j++, (tcd) = &(*trace_data[i])[j].tcd)
|
||||
|
||||
#define tcd_for_each_type_lock(tcd, i) \
|
||||
#define tcd_for_each_type_lock(tcd, i, cpu) \
|
||||
for (i = 0; trace_data[i] && \
|
||||
(tcd = &(*trace_data[i])[smp_processor_id()].tcd) && \
|
||||
(tcd = &(*trace_data[i])[cpu].tcd) && \
|
||||
trace_lock_tcd(tcd); trace_unlock_tcd(tcd), i++)
|
||||
|
||||
struct trace_page {
|
||||
|
@ -40,6 +40,7 @@
|
||||
#include <linux/kthread.h>
|
||||
#include <linux/hardirq.h>
|
||||
#include <linux/interrupt.h>
|
||||
#include <linux/spinlock.h>
|
||||
#include <sys/sysmacros.h>
|
||||
#include <sys/proc.h>
|
||||
#include <sys/debug.h>
|
||||
@ -424,35 +425,12 @@ trace_put_console_buffer(char *buffer)
|
||||
put_cpu();
|
||||
}
|
||||
|
||||
static struct trace_cpu_data *
|
||||
trace_get_tcd(void)
|
||||
{
|
||||
int cpu;
|
||||
|
||||
cpu = get_cpu();
|
||||
if (in_irq())
|
||||
return &(*trace_data[TCD_TYPE_IRQ])[cpu].tcd;
|
||||
else if (in_softirq())
|
||||
return &(*trace_data[TCD_TYPE_SOFTIRQ])[cpu].tcd;
|
||||
|
||||
return &(*trace_data[TCD_TYPE_PROC])[cpu].tcd;
|
||||
}
|
||||
|
||||
static void
|
||||
trace_put_tcd (struct trace_cpu_data *tcd)
|
||||
{
|
||||
put_cpu();
|
||||
}
|
||||
|
||||
static int
|
||||
trace_lock_tcd(struct trace_cpu_data *tcd)
|
||||
{
|
||||
__ASSERT(tcd->tcd_type < TCD_TYPE_MAX);
|
||||
|
||||
if (tcd->tcd_type == TCD_TYPE_IRQ)
|
||||
local_irq_disable();
|
||||
else if (tcd->tcd_type == TCD_TYPE_SOFTIRQ)
|
||||
local_bh_disable();
|
||||
spin_lock_irqsave(&tcd->tcd_lock, tcd->tcd_lock_flags);
|
||||
|
||||
return 1;
|
||||
}
|
||||
@ -462,10 +440,34 @@ trace_unlock_tcd(struct trace_cpu_data *tcd)
|
||||
{
|
||||
__ASSERT(tcd->tcd_type < TCD_TYPE_MAX);
|
||||
|
||||
if (tcd->tcd_type == TCD_TYPE_IRQ)
|
||||
local_irq_enable();
|
||||
else if (tcd->tcd_type == TCD_TYPE_SOFTIRQ)
|
||||
local_bh_enable();
|
||||
spin_unlock_irqrestore(&tcd->tcd_lock, tcd->tcd_lock_flags);
|
||||
}
|
||||
|
||||
static struct trace_cpu_data *
|
||||
trace_get_tcd(void)
|
||||
{
|
||||
int cpu;
|
||||
struct trace_cpu_data *tcd;
|
||||
|
||||
cpu = get_cpu();
|
||||
if (in_irq())
|
||||
tcd = &(*trace_data[TCD_TYPE_IRQ])[cpu].tcd;
|
||||
else if (in_softirq())
|
||||
tcd = &(*trace_data[TCD_TYPE_SOFTIRQ])[cpu].tcd;
|
||||
else
|
||||
tcd = &(*trace_data[TCD_TYPE_PROC])[cpu].tcd;
|
||||
|
||||
trace_lock_tcd(tcd);
|
||||
|
||||
return tcd;
|
||||
}
|
||||
|
||||
static void
|
||||
trace_put_tcd (struct trace_cpu_data *tcd)
|
||||
{
|
||||
trace_unlock_tcd(tcd);
|
||||
|
||||
put_cpu();
|
||||
}
|
||||
|
||||
static void
|
||||
@ -525,23 +527,6 @@ trace_max_debug_mb(void)
|
||||
return MAX(512, ((num_physpages >> (20 - PAGE_SHIFT)) * 80) / 100);
|
||||
}
|
||||
|
||||
static void
|
||||
trace_call_on_all_cpus(void (*fn)(void *arg), void *arg)
|
||||
{
|
||||
cpumask_t mask, cpus_allowed = current->cpus_allowed;
|
||||
int cpu;
|
||||
|
||||
for_each_online_cpu(cpu) {
|
||||
cpus_clear(mask);
|
||||
cpu_set(cpu, mask);
|
||||
set_cpus_allowed(current, mask);
|
||||
|
||||
fn(arg);
|
||||
|
||||
set_cpus_allowed(current, cpus_allowed);
|
||||
}
|
||||
}
|
||||
|
||||
static struct trace_page *
|
||||
tage_alloc(int gfp)
|
||||
{
|
||||
@ -861,16 +846,17 @@ collect_pages_from_single_cpu(struct page_collection *pc)
|
||||
}
|
||||
|
||||
static void
|
||||
collect_pages_on_cpu(void *info)
|
||||
collect_pages_on_all_cpus(struct page_collection *pc)
|
||||
{
|
||||
struct trace_cpu_data *tcd;
|
||||
struct page_collection *pc = info;
|
||||
int i;
|
||||
int i, cpu;
|
||||
|
||||
spin_lock(&pc->pc_lock);
|
||||
tcd_for_each_type_lock(tcd, i) {
|
||||
list_splice_init(&tcd->tcd_pages, &pc->pc_pages);
|
||||
tcd->tcd_cur_pages = 0;
|
||||
for_each_possible_cpu(cpu) {
|
||||
tcd_for_each_type_lock(tcd, i, cpu) {
|
||||
list_splice_init(&tcd->tcd_pages, &pc->pc_pages);
|
||||
tcd->tcd_cur_pages = 0;
|
||||
}
|
||||
}
|
||||
spin_unlock(&pc->pc_lock);
|
||||
}
|
||||
@ -883,34 +869,38 @@ collect_pages(dumplog_priv_t *dp, struct page_collection *pc)
|
||||
if (spl_panic_in_progress || dp->dp_flags & DL_SINGLE_CPU)
|
||||
collect_pages_from_single_cpu(pc);
|
||||
else
|
||||
trace_call_on_all_cpus(collect_pages_on_cpu, pc);
|
||||
collect_pages_on_all_cpus(pc);
|
||||
}
|
||||
|
||||
static void
|
||||
put_pages_back_on_cpu(void *info)
|
||||
put_pages_back_on_all_cpus(struct page_collection *pc)
|
||||
{
|
||||
struct page_collection *pc = info;
|
||||
struct trace_cpu_data *tcd;
|
||||
struct list_head *cur_head;
|
||||
struct trace_page *tage;
|
||||
struct trace_page *tmp;
|
||||
int i;
|
||||
int i, cpu;
|
||||
|
||||
spin_lock(&pc->pc_lock);
|
||||
tcd_for_each_type_lock(tcd, i) {
|
||||
cur_head = tcd->tcd_pages.next;
|
||||
|
||||
list_for_each_entry_safe(tage, tmp, &pc->pc_pages, linkage) {
|
||||
for_each_possible_cpu(cpu) {
|
||||
tcd_for_each_type_lock(tcd, i, cpu) {
|
||||
cur_head = tcd->tcd_pages.next;
|
||||
|
||||
__ASSERT_TAGE_INVARIANT(tage);
|
||||
list_for_each_entry_safe(tage, tmp, &pc->pc_pages,
|
||||
linkage) {
|
||||
|
||||
if (tage->cpu != smp_processor_id() || tage->type != i)
|
||||
continue;
|
||||
__ASSERT_TAGE_INVARIANT(tage);
|
||||
|
||||
tage_to_tail(tage, cur_head);
|
||||
tcd->tcd_cur_pages++;
|
||||
if (tage->cpu != cpu || tage->type != i)
|
||||
continue;
|
||||
|
||||
tage_to_tail(tage, cur_head);
|
||||
tcd->tcd_cur_pages++;
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
spin_unlock(&pc->pc_lock);
|
||||
}
|
||||
|
||||
@ -918,7 +908,7 @@ static void
|
||||
put_pages_back(struct page_collection *pc)
|
||||
{
|
||||
if (!spl_panic_in_progress)
|
||||
trace_call_on_all_cpus(put_pages_back_on_cpu, pc);
|
||||
put_pages_back_on_all_cpus(pc);
|
||||
}
|
||||
|
||||
static struct file *
|
||||
@ -1177,6 +1167,7 @@ trace_init(int max_pages)
|
||||
}
|
||||
|
||||
tcd_for_each(tcd, i, j) {
|
||||
spin_lock_init(&tcd->tcd_lock);
|
||||
tcd->tcd_pages_factor = pages_factor[i];
|
||||
tcd->tcd_type = i;
|
||||
tcd->tcd_cpu = j;
|
||||
@ -1231,23 +1222,26 @@ debug_init(void)
|
||||
}
|
||||
|
||||
static void
|
||||
trace_cleanup_on_cpu(void *info)
|
||||
trace_cleanup_on_all_cpus(void)
|
||||
{
|
||||
struct trace_cpu_data *tcd;
|
||||
struct trace_page *tage;
|
||||
struct trace_page *tmp;
|
||||
int i;
|
||||
int i, cpu;
|
||||
|
||||
tcd_for_each_type_lock(tcd, i) {
|
||||
tcd->tcd_shutting_down = 1;
|
||||
for_each_possible_cpu(cpu) {
|
||||
tcd_for_each_type_lock(tcd, i, cpu) {
|
||||
tcd->tcd_shutting_down = 1;
|
||||
|
||||
list_for_each_entry_safe(tage, tmp, &tcd->tcd_pages, linkage) {
|
||||
__ASSERT_TAGE_INVARIANT(tage);
|
||||
list_for_each_entry_safe(tage, tmp, &tcd->tcd_pages,
|
||||
linkage) {
|
||||
__ASSERT_TAGE_INVARIANT(tage);
|
||||
|
||||
list_del(&tage->linkage);
|
||||
tage_free(tage);
|
||||
list_del(&tage->linkage);
|
||||
tage_free(tage);
|
||||
}
|
||||
tcd->tcd_cur_pages = 0;
|
||||
}
|
||||
tcd->tcd_cur_pages = 0;
|
||||
}
|
||||
}
|
||||
|
||||
@ -1256,7 +1250,7 @@ trace_fini(void)
|
||||
{
|
||||
int i, j;
|
||||
|
||||
trace_call_on_all_cpus(trace_cleanup_on_cpu, NULL);
|
||||
trace_cleanup_on_all_cpus();
|
||||
|
||||
for (i = 0; i < num_possible_cpus(); i++) {
|
||||
for (j = 0; j < 3; j++) {
|
||||
|
Loading…
Reference in New Issue
Block a user