2005-07-14 17:40:02 +00:00
|
|
|
/*-
|
|
|
|
* Copyright (c) 2005 Robert N. M. Watson
|
|
|
|
* All rights reserved.
|
|
|
|
*
|
|
|
|
* Redistribution and use in source and binary forms, with or without
|
|
|
|
* modification, are permitted provided that the following conditions
|
|
|
|
* are met:
|
|
|
|
* 1. Redistributions of source code must retain the above copyright
|
|
|
|
* notice, this list of conditions and the following disclaimer.
|
|
|
|
* 2. Redistributions in binary form must reproduce the above copyright
|
|
|
|
* notice, this list of conditions and the following disclaimer in the
|
|
|
|
* documentation and/or other materials provided with the distribution.
|
|
|
|
*
|
|
|
|
* THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND
|
|
|
|
* ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
|
|
|
|
* IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
|
|
|
|
* ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE
|
|
|
|
* FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
|
|
|
|
* DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
|
|
|
|
* OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
|
|
|
|
* HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
|
|
|
|
* LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
|
|
|
|
* OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
|
|
|
|
* SUCH DAMAGE.
|
|
|
|
*
|
|
|
|
* $FreeBSD$
|
|
|
|
*/
|
|
|
|
|
|
|
|
#include <sys/param.h>
|
|
|
|
#include <sys/sysctl.h>
|
|
|
|
|
|
|
|
#include <err.h>
|
|
|
|
#include <errno.h>
|
|
|
|
#include <stdio.h>
|
|
|
|
#include <stdlib.h>
|
|
|
|
#include <string.h>
|
|
|
|
|
|
|
|
#include "memstat.h"
|
|
|
|
#include "memstat_internal.h"
|
|
|
|
|
2005-07-24 01:41:47 +00:00
|
|
|
const char *
|
|
|
|
memstat_strerror(int error)
|
|
|
|
{
|
|
|
|
|
|
|
|
switch (error) {
|
|
|
|
case MEMSTAT_ERROR_NOMEMORY:
|
|
|
|
return ("Cannot allocate memory");
|
|
|
|
case MEMSTAT_ERROR_VERSION:
|
|
|
|
return ("Version mismatch");
|
|
|
|
case MEMSTAT_ERROR_PERMISSION:
|
|
|
|
return ("Permission denied");
|
|
|
|
case MEMSTAT_ERROR_TOOMANYCPUS:
|
|
|
|
return ("Too many CPUs");
|
|
|
|
case MEMSTAT_ERROR_DATAERROR:
|
|
|
|
return ("Data format error");
|
2005-08-01 19:07:39 +00:00
|
|
|
case MEMSTAT_ERROR_KVM:
|
|
|
|
return ("KVM error");
|
|
|
|
case MEMSTAT_ERROR_KVM_NOSYMBOL:
|
|
|
|
return ("KVM unable to find symbol");
|
|
|
|
case MEMSTAT_ERROR_KVM_SHORTREAD:
|
|
|
|
return ("KVM short read");
|
2005-07-24 01:41:47 +00:00
|
|
|
case MEMSTAT_ERROR_UNDEFINED:
|
|
|
|
default:
|
|
|
|
return ("Unknown error");
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2005-07-14 17:40:02 +00:00
|
|
|
struct memory_type_list *
|
|
|
|
memstat_mtl_alloc(void)
|
|
|
|
{
|
|
|
|
struct memory_type_list *mtlp;
|
|
|
|
|
|
|
|
mtlp = malloc(sizeof(*mtlp));
|
|
|
|
if (mtlp == NULL)
|
|
|
|
return (NULL);
|
|
|
|
|
2005-07-24 01:28:54 +00:00
|
|
|
LIST_INIT(&mtlp->mtl_list);
|
|
|
|
mtlp->mtl_error = MEMSTAT_ERROR_UNDEFINED;
|
2005-07-14 17:40:02 +00:00
|
|
|
return (mtlp);
|
|
|
|
}
|
|
|
|
|
|
|
|
struct memory_type *
|
|
|
|
memstat_mtl_first(struct memory_type_list *list)
|
|
|
|
{
|
|
|
|
|
2005-07-24 01:28:54 +00:00
|
|
|
return (LIST_FIRST(&list->mtl_list));
|
2005-07-14 17:40:02 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
struct memory_type *
|
|
|
|
memstat_mtl_next(struct memory_type *mtp)
|
|
|
|
{
|
|
|
|
|
|
|
|
return (LIST_NEXT(mtp, mt_list));
|
|
|
|
}
|
|
|
|
|
|
|
|
void
|
2005-08-01 13:18:21 +00:00
|
|
|
_memstat_mtl_empty(struct memory_type_list *list)
|
2005-07-14 17:40:02 +00:00
|
|
|
{
|
|
|
|
struct memory_type *mtp;
|
|
|
|
|
2005-07-24 01:28:54 +00:00
|
|
|
while ((mtp = LIST_FIRST(&list->mtl_list))) {
|
2005-07-14 17:40:02 +00:00
|
|
|
LIST_REMOVE(mtp, mt_list);
|
|
|
|
free(mtp);
|
|
|
|
}
|
2005-08-01 13:18:21 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
void
|
|
|
|
memstat_mtl_free(struct memory_type_list *list)
|
|
|
|
{
|
|
|
|
|
|
|
|
_memstat_mtl_empty(list);
|
2005-07-14 17:40:02 +00:00
|
|
|
free(list);
|
|
|
|
}
|
|
|
|
|
2005-07-24 01:28:54 +00:00
|
|
|
int
|
|
|
|
memstat_mtl_geterror(struct memory_type_list *list)
|
|
|
|
{
|
|
|
|
|
|
|
|
return (list->mtl_error);
|
|
|
|
}
|
|
|
|
|
2005-07-14 17:40:02 +00:00
|
|
|
/*
|
|
|
|
* Look for an existing memory_type entry in a memory_type list, based on the
|
2005-07-24 01:28:54 +00:00
|
|
|
* allocator and name of the type. If not found, return NULL. No errno or
|
|
|
|
* memstat error.
|
2005-07-14 17:40:02 +00:00
|
|
|
*/
|
|
|
|
struct memory_type *
|
|
|
|
memstat_mtl_find(struct memory_type_list *list, int allocator,
|
|
|
|
const char *name)
|
|
|
|
{
|
|
|
|
struct memory_type *mtp;
|
|
|
|
|
2005-07-24 01:28:54 +00:00
|
|
|
LIST_FOREACH(mtp, &list->mtl_list, mt_list) {
|
2005-07-14 17:40:02 +00:00
|
|
|
if ((mtp->mt_allocator == allocator ||
|
|
|
|
allocator == ALLOCATOR_ANY) &&
|
|
|
|
strcmp(mtp->mt_name, name) == 0)
|
|
|
|
return (mtp);
|
|
|
|
}
|
|
|
|
return (NULL);
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Allocate a new memory_type with the specificed allocator type and name,
|
|
|
|
* then insert into the list. The structure will be zero'd.
|
2005-07-23 21:17:15 +00:00
|
|
|
*
|
|
|
|
* libmemstat(3) internal function.
|
2005-07-14 17:40:02 +00:00
|
|
|
*/
|
|
|
|
struct memory_type *
|
2005-07-23 21:17:15 +00:00
|
|
|
_memstat_mt_allocate(struct memory_type_list *list, int allocator,
|
2005-07-14 17:40:02 +00:00
|
|
|
const char *name)
|
|
|
|
{
|
|
|
|
struct memory_type *mtp;
|
|
|
|
|
|
|
|
mtp = malloc(sizeof(*mtp));
|
|
|
|
if (mtp == NULL)
|
|
|
|
return (NULL);
|
|
|
|
|
|
|
|
bzero(mtp, sizeof(*mtp));
|
|
|
|
|
|
|
|
mtp->mt_allocator = allocator;
|
|
|
|
strlcpy(mtp->mt_name, name, MEMTYPE_MAXNAME);
|
2005-07-24 01:28:54 +00:00
|
|
|
LIST_INSERT_HEAD(&list->mtl_list, mtp, mt_list);
|
2005-07-14 17:40:02 +00:00
|
|
|
return (mtp);
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Reset any libmemstat(3)-owned statistics in a memory_type record so that
|
|
|
|
* it can be reused without incremental addition problems. Caller-owned
|
|
|
|
* memory is left "as-is", and must be updated by the caller if desired.
|
2005-07-23 21:17:15 +00:00
|
|
|
*
|
|
|
|
* libmemstat(3) internal function.
|
2005-07-14 17:40:02 +00:00
|
|
|
*/
|
|
|
|
void
|
2005-07-23 21:17:15 +00:00
|
|
|
_memstat_mt_reset_stats(struct memory_type *mtp)
|
2005-07-14 17:40:02 +00:00
|
|
|
{
|
|
|
|
int i;
|
|
|
|
|
|
|
|
mtp->mt_countlimit = 0;
|
|
|
|
mtp->mt_byteslimit = 0;
|
|
|
|
mtp->mt_sizemask = 0;
|
|
|
|
mtp->mt_size = 0;
|
|
|
|
|
|
|
|
mtp->mt_memalloced = 0;
|
|
|
|
mtp->mt_memfreed = 0;
|
|
|
|
mtp->mt_numallocs = 0;
|
|
|
|
mtp->mt_numfrees = 0;
|
|
|
|
mtp->mt_bytes = 0;
|
|
|
|
mtp->mt_count = 0;
|
|
|
|
mtp->mt_free = 0;
|
|
|
|
mtp->mt_failures = 0;
|
2010-06-15 19:28:37 +00:00
|
|
|
mtp->mt_sleeps = 0;
|
2005-07-14 17:40:02 +00:00
|
|
|
|
|
|
|
mtp->mt_zonefree = 0;
|
UMA supports "secondary" zones, in which a second zone can be layered
on top of a primary zone, sharing the same allocation "keg". When
reporting statistics for zones, do not report the free items in the
keg as part of the free items in the zone, or those free items will
be reported more than once: for the primary zone, and then any
secondary zones off the primary zone. Separately record and maintain
a kegfree statistic, and export via memstat_get_kegfree(), which is
available for use if needed. Since items free'd back to the keg are
not fully initialized, and hence may not actually be available (since
secondary zone ctor-time initialization can fail), this makes some
amount of sense.
This change corrects a bug made visible in the libmemstat(3)
modifications to netstat: mbufs freed back to the keg from the
packet zone would be counted twice, resulting in negative values
being printed in the mbuf free count.
Some further refinement of reporting relating to secondary zones may
still be required.
Reported by: ssouhlal
MFC after: 3 days
2005-07-20 09:17:40 +00:00
|
|
|
mtp->mt_kegfree = 0;
|
2005-07-14 17:40:02 +00:00
|
|
|
|
|
|
|
for (i = 0; i < MEMSTAT_MAXCPU; i++) {
|
|
|
|
mtp->mt_percpu_alloc[i].mtp_memalloced = 0;
|
|
|
|
mtp->mt_percpu_alloc[i].mtp_memfreed = 0;
|
|
|
|
mtp->mt_percpu_alloc[i].mtp_numallocs = 0;
|
|
|
|
mtp->mt_percpu_alloc[i].mtp_numfrees = 0;
|
|
|
|
mtp->mt_percpu_alloc[i].mtp_sizemask = 0;
|
|
|
|
mtp->mt_percpu_cache[i].mtp_free = 0;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Accessor methods for struct memory_type. Avoids encoding the structure
|
|
|
|
* ABI into the application.
|
|
|
|
*/
|
|
|
|
const char *
|
|
|
|
memstat_get_name(const struct memory_type *mtp)
|
|
|
|
{
|
|
|
|
|
|
|
|
return (mtp->mt_name);
|
|
|
|
}
|
|
|
|
|
|
|
|
int
|
|
|
|
memstat_get_allocator(const struct memory_type *mtp)
|
|
|
|
{
|
|
|
|
|
|
|
|
return (mtp->mt_allocator);
|
|
|
|
}
|
|
|
|
|
|
|
|
uint64_t
|
|
|
|
memstat_get_countlimit(const struct memory_type *mtp)
|
|
|
|
{
|
|
|
|
|
|
|
|
return (mtp->mt_countlimit);
|
|
|
|
}
|
|
|
|
|
|
|
|
uint64_t
|
|
|
|
memstat_get_byteslimit(const struct memory_type *mtp)
|
|
|
|
{
|
|
|
|
|
|
|
|
return (mtp->mt_byteslimit);
|
|
|
|
}
|
|
|
|
|
|
|
|
uint64_t
|
|
|
|
memstat_get_sizemask(const struct memory_type *mtp)
|
|
|
|
{
|
|
|
|
|
|
|
|
return (mtp->mt_sizemask);
|
|
|
|
}
|
|
|
|
|
|
|
|
uint64_t
|
|
|
|
memstat_get_size(const struct memory_type *mtp)
|
|
|
|
{
|
|
|
|
|
|
|
|
return (mtp->mt_size);
|
|
|
|
}
|
|
|
|
|
|
|
|
uint64_t
|
|
|
|
memstat_get_memalloced(const struct memory_type *mtp)
|
|
|
|
{
|
|
|
|
|
|
|
|
return (mtp->mt_memalloced);
|
|
|
|
}
|
|
|
|
|
|
|
|
uint64_t
|
|
|
|
memstat_get_memfreed(const struct memory_type *mtp)
|
|
|
|
{
|
|
|
|
|
|
|
|
return (mtp->mt_memfreed);
|
|
|
|
}
|
|
|
|
|
|
|
|
uint64_t
|
|
|
|
memstat_get_numallocs(const struct memory_type *mtp)
|
|
|
|
{
|
|
|
|
|
|
|
|
return (mtp->mt_numallocs);
|
|
|
|
}
|
|
|
|
|
|
|
|
uint64_t
|
|
|
|
memstat_get_numfrees(const struct memory_type *mtp)
|
|
|
|
{
|
|
|
|
|
|
|
|
return (mtp->mt_numfrees);
|
|
|
|
}
|
|
|
|
|
|
|
|
uint64_t
|
|
|
|
memstat_get_bytes(const struct memory_type *mtp)
|
|
|
|
{
|
|
|
|
|
|
|
|
return (mtp->mt_bytes);
|
|
|
|
}
|
|
|
|
|
|
|
|
uint64_t
|
|
|
|
memstat_get_count(const struct memory_type *mtp)
|
|
|
|
{
|
|
|
|
|
|
|
|
return (mtp->mt_count);
|
|
|
|
}
|
|
|
|
|
|
|
|
uint64_t
|
|
|
|
memstat_get_free(const struct memory_type *mtp)
|
|
|
|
{
|
|
|
|
|
|
|
|
return (mtp->mt_free);
|
|
|
|
}
|
|
|
|
|
|
|
|
uint64_t
|
|
|
|
memstat_get_failures(const struct memory_type *mtp)
|
|
|
|
{
|
|
|
|
|
|
|
|
return (mtp->mt_failures);
|
|
|
|
}
|
|
|
|
|
2010-06-15 19:28:37 +00:00
|
|
|
uint64_t
|
|
|
|
memstat_get_sleeps(const struct memory_type *mtp)
|
|
|
|
{
|
|
|
|
|
|
|
|
return (mtp->mt_sleeps);
|
|
|
|
}
|
|
|
|
|
2005-07-14 17:40:02 +00:00
|
|
|
void *
|
|
|
|
memstat_get_caller_pointer(const struct memory_type *mtp, int index)
|
|
|
|
{
|
|
|
|
|
|
|
|
return (mtp->mt_caller_pointer[index]);
|
|
|
|
}
|
|
|
|
|
|
|
|
void
|
|
|
|
memstat_set_caller_pointer(struct memory_type *mtp, int index, void *value)
|
|
|
|
{
|
|
|
|
|
|
|
|
mtp->mt_caller_pointer[index] = value;
|
|
|
|
}
|
|
|
|
|
|
|
|
uint64_t
|
|
|
|
memstat_get_caller_uint64(const struct memory_type *mtp, int index)
|
|
|
|
{
|
|
|
|
|
|
|
|
return (mtp->mt_caller_uint64[index]);
|
|
|
|
}
|
|
|
|
|
|
|
|
void
|
|
|
|
memstat_set_caller_uint64(struct memory_type *mtp, int index, uint64_t value)
|
|
|
|
{
|
|
|
|
|
|
|
|
mtp->mt_caller_uint64[index] = value;
|
|
|
|
}
|
|
|
|
|
|
|
|
uint64_t
|
|
|
|
memstat_get_zonefree(const struct memory_type *mtp)
|
|
|
|
{
|
|
|
|
|
|
|
|
return (mtp->mt_zonefree);
|
|
|
|
}
|
|
|
|
|
UMA supports "secondary" zones, in which a second zone can be layered
on top of a primary zone, sharing the same allocation "keg". When
reporting statistics for zones, do not report the free items in the
keg as part of the free items in the zone, or those free items will
be reported more than once: for the primary zone, and then any
secondary zones off the primary zone. Separately record and maintain
a kegfree statistic, and export via memstat_get_kegfree(), which is
available for use if needed. Since items free'd back to the keg are
not fully initialized, and hence may not actually be available (since
secondary zone ctor-time initialization can fail), this makes some
amount of sense.
This change corrects a bug made visible in the libmemstat(3)
modifications to netstat: mbufs freed back to the keg from the
packet zone would be counted twice, resulting in negative values
being printed in the mbuf free count.
Some further refinement of reporting relating to secondary zones may
still be required.
Reported by: ssouhlal
MFC after: 3 days
2005-07-20 09:17:40 +00:00
|
|
|
uint64_t
|
|
|
|
memstat_get_kegfree(const struct memory_type *mtp)
|
|
|
|
{
|
|
|
|
|
|
|
|
return (mtp->mt_kegfree);
|
|
|
|
}
|
|
|
|
|
2005-07-14 17:40:02 +00:00
|
|
|
uint64_t
|
|
|
|
memstat_get_percpu_memalloced(const struct memory_type *mtp, int cpu)
|
|
|
|
{
|
|
|
|
|
|
|
|
return (mtp->mt_percpu_alloc[cpu].mtp_memalloced);
|
|
|
|
}
|
|
|
|
|
|
|
|
uint64_t
|
|
|
|
memstat_get_percpu_memfreed(const struct memory_type *mtp, int cpu)
|
|
|
|
{
|
|
|
|
|
|
|
|
return (mtp->mt_percpu_alloc[cpu].mtp_memfreed);
|
|
|
|
}
|
|
|
|
|
|
|
|
uint64_t
|
|
|
|
memstat_get_percpu_numallocs(const struct memory_type *mtp, int cpu)
|
|
|
|
{
|
|
|
|
|
|
|
|
return (mtp->mt_percpu_alloc[cpu].mtp_numallocs);
|
|
|
|
}
|
|
|
|
|
|
|
|
uint64_t
|
|
|
|
memstat_get_percpu_numfrees(const struct memory_type *mtp, int cpu)
|
|
|
|
{
|
|
|
|
|
|
|
|
return (mtp->mt_percpu_alloc[cpu].mtp_numfrees);
|
|
|
|
}
|
|
|
|
|
|
|
|
uint64_t
|
|
|
|
memstat_get_percpu_sizemask(const struct memory_type *mtp, int cpu)
|
|
|
|
{
|
|
|
|
|
|
|
|
return (mtp->mt_percpu_alloc[cpu].mtp_sizemask);
|
|
|
|
}
|
|
|
|
|
|
|
|
void *
|
|
|
|
memstat_get_percpu_caller_pointer(const struct memory_type *mtp, int cpu,
|
|
|
|
int index)
|
|
|
|
{
|
|
|
|
|
|
|
|
return (mtp->mt_percpu_alloc[cpu].mtp_caller_pointer[index]);
|
|
|
|
}
|
|
|
|
|
|
|
|
void
|
|
|
|
memstat_set_percpu_caller_pointer(struct memory_type *mtp, int cpu,
|
|
|
|
int index, void *value)
|
|
|
|
{
|
|
|
|
|
|
|
|
mtp->mt_percpu_alloc[cpu].mtp_caller_pointer[index] = value;
|
|
|
|
}
|
|
|
|
|
|
|
|
uint64_t
|
|
|
|
memstat_get_percpu_caller_uint64(const struct memory_type *mtp, int cpu,
|
|
|
|
int index)
|
|
|
|
{
|
|
|
|
|
|
|
|
return (mtp->mt_percpu_alloc[cpu].mtp_caller_uint64[index]);
|
|
|
|
}
|
|
|
|
|
|
|
|
void
|
|
|
|
memstat_set_percpu_caller_uint64(struct memory_type *mtp, int cpu, int index,
|
|
|
|
uint64_t value)
|
|
|
|
{
|
|
|
|
|
|
|
|
mtp->mt_percpu_alloc[cpu].mtp_caller_uint64[index] = value;
|
|
|
|
}
|
|
|
|
|
|
|
|
uint64_t
|
|
|
|
memstat_get_percpu_free(const struct memory_type *mtp, int cpu)
|
|
|
|
{
|
|
|
|
|
|
|
|
return (mtp->mt_percpu_cache[cpu].mtp_free);
|
|
|
|
}
|