freebsd-nq/sys/dev/acpica/Osd/OsdSchedule.c
Nate Lawson 4c12fb6750 Fix GPE livelock that occurs on HP/Compaq laptops, mostly in the thermal
zone code.  The GPE handler method (i.e. _L00) generates various Notify
events that need to be run to completion before the GPE is re-enabled.
In ACPI-CA, we queue an asynch callback at the same priority as a Notify
so that it will only run after all Notify handlers have completed.  The
callback re-enables the GPE afterwards.  We also changed the priority of
Notifies to be the same as GPEs, given the possibility that another GPE
could arrive before the Notifies have completed and we don't want it to
get queued ahead of the rest.

The ACPI-CA change was submitted by Alexey Starikovskiy (SUSE) and will
appear in a later release.  Special thanks to him for helping track this
bug down.

MFC after:	1 week
Tested by:	jhb, Yousif Hassan <yousif / alumni.jmu.edu>
2008-01-12 22:13:12 +00:00

200 lines
5.2 KiB
C

/*-
* Copyright (c) 2000 Michael Smith
* Copyright (c) 2000 BSDi
* All rights reserved.
*
* Redistribution and use in source and binary forms, with or without
* modification, are permitted provided that the following conditions
* are met:
* 1. Redistributions of source code must retain the above copyright
* notice, this list of conditions and the following disclaimer.
* 2. Redistributions in binary form must reproduce the above copyright
* notice, this list of conditions and the following disclaimer in the
* documentation and/or other materials provided with the distribution.
*
* THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND
* ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
* IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
* ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE
* FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
* DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
* OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
* HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
* LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
* OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
* SUCH DAMAGE.
*/
/*
* 6.3 : Scheduling services
*/
#include <sys/cdefs.h>
__FBSDID("$FreeBSD$");
#include "opt_acpi.h"
#include <sys/param.h>
#include <sys/systm.h>
#include <sys/bus.h>
#include <sys/interrupt.h>
#include <sys/kernel.h>
#include <sys/kthread.h>
#include <sys/malloc.h>
#include <sys/proc.h>
#include <sys/taskqueue.h>
#include <contrib/dev/acpica/acpi.h>
#include <dev/acpica/acpivar.h>
#define _COMPONENT ACPI_OS_SERVICES
ACPI_MODULE_NAME("SCHEDULE")
/*
* Allow the user to tune the number of task threads we start. It seems
* some systems have problems with increased parallelism.
*/
static int acpi_max_threads = ACPI_MAX_THREADS;
TUNABLE_INT("debug.acpi.max_threads", &acpi_max_threads);
MALLOC_DEFINE(M_ACPITASK, "acpitask", "ACPI deferred task");
struct acpi_task_ctx {
struct task at_task;
ACPI_OSD_EXEC_CALLBACK at_function;
void *at_context;
};
TASKQUEUE_DEFINE(acpi, taskqueue_thread_enqueue, &taskqueue_acpi,
taskqueue_start_threads(&taskqueue_acpi, acpi_max_threads, PWAIT,
"acpi_task"));
/*
* Bounce through this wrapper function since ACPI-CA doesn't understand
* the pending argument for its callbacks.
*/
static void
acpi_task_execute(void *context, int pending)
{
struct acpi_task_ctx *at;
at = (struct acpi_task_ctx *)context;
at->at_function(at->at_context);
free(at, M_ACPITASK);
}
/*
* This function may be called in interrupt context, i.e. when a GPE fires.
* We allocate and queue a task for one of our taskqueue threads to process.
*/
ACPI_STATUS
AcpiOsExecute(ACPI_EXECUTE_TYPE Type, ACPI_OSD_EXEC_CALLBACK Function,
void *Context)
{
struct acpi_task_ctx *at;
int pri;
ACPI_FUNCTION_TRACE((char *)(uintptr_t)__func__);
if (Function == NULL)
return_ACPI_STATUS (AE_BAD_PARAMETER);
at = malloc(sizeof(*at), M_ACPITASK, M_NOWAIT);
if (at == NULL)
return_ACPI_STATUS (AE_NO_MEMORY);
at->at_function = Function;
at->at_context = Context;
switch (Type) {
case OSL_GPE_HANDLER:
case OSL_NOTIFY_HANDLER:
/*
* Run GPEs and Notifies at the same priority. This allows
* Notifies that are generated by running a GPE's method (e.g., _L00)
* to not be pre-empted by a later GPE that arrives during the
* Notify handler execution.
*/
pri = 10;
break;
case OSL_GLOBAL_LOCK_HANDLER:
case OSL_EC_POLL_HANDLER:
case OSL_EC_BURST_HANDLER:
pri = 5;
break;
case OSL_DEBUGGER_THREAD:
pri = 0;
break;
default:
free(at, M_ACPITASK);
return_ACPI_STATUS (AE_BAD_PARAMETER);
}
TASK_INIT(&at->at_task, pri, acpi_task_execute, at);
taskqueue_enqueue(taskqueue_acpi, &at->at_task);
return_ACPI_STATUS (AE_OK);
}
void
AcpiOsSleep(ACPI_INTEGER Milliseconds)
{
int timo;
ACPI_FUNCTION_TRACE((char *)(uintptr_t)__func__);
timo = Milliseconds * hz / 1000;
/*
* If requested sleep time is less than our hz resolution, use
* DELAY instead for better granularity.
*/
if (timo > 0)
pause("acpislp", timo);
else
DELAY(Milliseconds * 1000);
return_VOID;
}
/*
* Return the current time in 100 nanosecond units
*/
UINT64
AcpiOsGetTimer(void)
{
struct bintime bt;
UINT64 t;
/* XXX During early boot there is no (decent) timer available yet. */
if (cold)
panic("acpi: timer op not yet supported during boot");
binuptime(&bt);
t = ((UINT64)10000000 * (uint32_t)(bt.frac >> 32)) >> 32;
t += bt.sec * 10000000;
return (t);
}
void
AcpiOsStall(UINT32 Microseconds)
{
ACPI_FUNCTION_TRACE((char *)(uintptr_t)__func__);
DELAY(Microseconds);
return_VOID;
}
ACPI_THREAD_ID
AcpiOsGetThreadId(void)
{
struct proc *p;
/* XXX do not add ACPI_FUNCTION_TRACE here, results in recursive call. */
p = curproc;
KASSERT(p != NULL, ("%s: curproc is NULL!", __func__));
/* Returning 0 is not allowed. */
return (p->p_pid + 1);
}