701 lines
14 KiB
C
701 lines
14 KiB
C
/* $OpenBSD: kstat.c,v 1.2 2022/01/31 05:09:17 dlg Exp $ */
|
|
|
|
/*
|
|
* Copyright (c) 2020 David Gwynne <dlg@openbsd.org>
|
|
*
|
|
* Permission to use, copy, modify, and distribute this software for any
|
|
* purpose with or without fee is hereby granted, provided that the above
|
|
* copyright notice and this permission notice appear in all copies.
|
|
*
|
|
* THE SOFTWARE IS PROVIDED "AS IS" AND THE AUTHOR DISCLAIMS ALL WARRANTIES
|
|
* WITH REGARD TO THIS SOFTWARE INCLUDING ALL IMPLIED WARRANTIES OF
|
|
* MERCHANTABILITY AND FITNESS. IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR
|
|
* ANY SPECIAL, DIRECT, INDIRECT, OR CONSEQUENTIAL DAMAGES OR ANY DAMAGES
|
|
* WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR PROFITS, WHETHER IN AN
|
|
* ACTION OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS ACTION, ARISING OUT OF
|
|
* OR IN CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOFTWARE.
|
|
*/
|
|
|
|
#include <sys/param.h>
|
|
#include <sys/systm.h>
|
|
#include <sys/types.h>
|
|
#include <sys/malloc.h>
|
|
#include <sys/pool.h>
|
|
#include <sys/time.h>
|
|
|
|
/* for kstat_set_cpu */
|
|
#include <sys/proc.h>
|
|
#include <sys/sched.h>
|
|
|
|
#include <sys/kstat.h>
|
|
|
|
RBT_HEAD(kstat_id_tree, kstat);
|
|
|
|
static inline int
|
|
kstat_id_cmp(const struct kstat *a, const struct kstat *b)
|
|
{
|
|
if (a->ks_id > b->ks_id)
|
|
return (1);
|
|
if (a->ks_id < b->ks_id)
|
|
return (-1);
|
|
|
|
return (0);
|
|
}
|
|
|
|
RBT_PROTOTYPE(kstat_id_tree, kstat, ks_id_entry, kstat_id_cmp);
|
|
|
|
RBT_HEAD(kstat_pv_tree, kstat);
|
|
|
|
static inline int
|
|
kstat_pv_cmp(const struct kstat *a, const struct kstat *b)
|
|
{
|
|
int rv;
|
|
|
|
rv = strcmp(a->ks_provider, b->ks_provider);
|
|
if (rv != 0)
|
|
return (rv);
|
|
|
|
if (a->ks_instance > b->ks_instance)
|
|
return (1);
|
|
if (a->ks_instance < b->ks_instance)
|
|
return (-1);
|
|
|
|
rv = strcmp(a->ks_name, b->ks_name);
|
|
if (rv != 0)
|
|
return (rv);
|
|
|
|
if (a->ks_unit > b->ks_unit)
|
|
return (1);
|
|
if (a->ks_unit < b->ks_unit)
|
|
return (-1);
|
|
|
|
return (0);
|
|
}
|
|
|
|
RBT_PROTOTYPE(kstat_pv_tree, kstat, ks_pv_entry, kstat_pv_cmp);
|
|
|
|
RBT_HEAD(kstat_nm_tree, kstat);
|
|
|
|
static inline int
|
|
kstat_nm_cmp(const struct kstat *a, const struct kstat *b)
|
|
{
|
|
int rv;
|
|
|
|
rv = strcmp(a->ks_name, b->ks_name);
|
|
if (rv != 0)
|
|
return (rv);
|
|
|
|
if (a->ks_unit > b->ks_unit)
|
|
return (1);
|
|
if (a->ks_unit < b->ks_unit)
|
|
return (-1);
|
|
|
|
rv = strcmp(a->ks_provider, b->ks_provider);
|
|
if (rv != 0)
|
|
return (rv);
|
|
|
|
if (a->ks_instance > b->ks_instance)
|
|
return (1);
|
|
if (a->ks_instance < b->ks_instance)
|
|
return (-1);
|
|
|
|
return (0);
|
|
}
|
|
|
|
RBT_PROTOTYPE(kstat_nm_tree, kstat, ks_nm_entry, kstat_nm_cmp);
|
|
|
|
struct kstat_lock_ops {
|
|
void (*enter)(void *);
|
|
void (*leave)(void *);
|
|
};
|
|
|
|
#define kstat_enter(_ks) (_ks)->ks_lock_ops->enter((_ks)->ks_lock)
|
|
#define kstat_leave(_ks) (_ks)->ks_lock_ops->leave((_ks)->ks_lock)
|
|
|
|
const struct kstat_lock_ops kstat_rlock_ops = {
|
|
(void (*)(void *))rw_enter_read,
|
|
(void (*)(void *))rw_exit_read,
|
|
};
|
|
|
|
const struct kstat_lock_ops kstat_wlock_ops = {
|
|
(void (*)(void *))rw_enter_write,
|
|
(void (*)(void *))rw_exit_write,
|
|
};
|
|
|
|
const struct kstat_lock_ops kstat_mutex_ops = {
|
|
(void (*)(void *))mtx_enter,
|
|
(void (*)(void *))mtx_leave,
|
|
};
|
|
|
|
void kstat_cpu_enter(void *);
|
|
void kstat_cpu_leave(void *);
|
|
|
|
const struct kstat_lock_ops kstat_cpu_ops = {
|
|
kstat_cpu_enter,
|
|
kstat_cpu_leave,
|
|
};
|
|
|
|
struct rwlock kstat_lock = RWLOCK_INITIALIZER("kstat");
|
|
|
|
/*
|
|
* The global state is versioned so changes to the set of kstats
|
|
* can be detected. This is an int so it can be read atomically on
|
|
* any arch, which is a ridiculous optimisation, really.
|
|
*/
|
|
unsigned int kstat_version = 0;
|
|
|
|
/*
|
|
* kstat structures have a unique identifier so they can be found
|
|
* quickly. Identifiers are 64bit in the hope that it won't wrap
|
|
* during the runtime of a system. The identifiers start at 1 so that
|
|
* 0 can be used as the first value for userland to iterate with.
|
|
*/
|
|
uint64_t kstat_next_id = 1;
|
|
|
|
struct kstat_id_tree kstat_id_tree = RBT_INITIALIZER();
|
|
struct kstat_pv_tree kstat_pv_tree = RBT_INITIALIZER();
|
|
struct kstat_nm_tree kstat_nm_tree = RBT_INITIALIZER();
|
|
struct pool kstat_pool;
|
|
|
|
struct rwlock kstat_default_lock = RWLOCK_INITIALIZER("kstatlk");
|
|
|
|
int kstat_read(struct kstat *);
|
|
int kstat_copy(struct kstat *, void *);
|
|
|
|
int
|
|
kstatattach(int num)
|
|
{
|
|
/* XXX install system stats here */
|
|
return (0);
|
|
}
|
|
|
|
int
|
|
kstatopen(dev_t dev, int flag, int mode, struct proc *p)
|
|
{
|
|
return (0);
|
|
}
|
|
|
|
int
|
|
kstatclose(dev_t dev, int flag, int mode, struct proc *p)
|
|
{
|
|
return (0);
|
|
}
|
|
|
|
int
|
|
kstatioc_enter(struct kstat_req *ksreq)
|
|
{
|
|
int error;
|
|
|
|
error = rw_enter(&kstat_lock, RW_READ | RW_INTR);
|
|
if (error != 0)
|
|
return (error);
|
|
|
|
if (!ISSET(ksreq->ks_rflags, KSTATIOC_F_IGNVER) &&
|
|
ksreq->ks_version != kstat_version) {
|
|
error = EINVAL;
|
|
goto error;
|
|
}
|
|
|
|
return (0);
|
|
|
|
error:
|
|
rw_exit(&kstat_lock);
|
|
return (error);
|
|
}
|
|
|
|
int
|
|
kstatioc_leave(struct kstat_req *ksreq, struct kstat *ks)
|
|
{
|
|
void *buf = NULL;
|
|
size_t klen = 0, ulen = 0;
|
|
struct timespec updated;
|
|
int error = 0;
|
|
|
|
if (ks == NULL) {
|
|
error = ENOENT;
|
|
goto error;
|
|
}
|
|
|
|
switch (ks->ks_state) {
|
|
case KSTAT_S_CREATED:
|
|
ksreq->ks_updated = ks->ks_created;
|
|
ksreq->ks_interval.tv_sec = 0;
|
|
ksreq->ks_interval.tv_nsec = 0;
|
|
ksreq->ks_datalen = 0;
|
|
ksreq->ks_dataver = 0;
|
|
break;
|
|
|
|
case KSTAT_S_INSTALLED:
|
|
ksreq->ks_dataver = ks->ks_dataver;
|
|
ksreq->ks_interval = ks->ks_interval;
|
|
|
|
if (ksreq->ks_data == NULL) {
|
|
/* userland doesn't want actual data, so shortcut */
|
|
kstat_enter(ks);
|
|
ksreq->ks_datalen = ks->ks_datalen;
|
|
ksreq->ks_updated = ks->ks_updated;
|
|
kstat_leave(ks);
|
|
break;
|
|
}
|
|
|
|
klen = ks->ks_datalen; /* KSTAT_F_REALLOC */
|
|
buf = malloc(klen, M_TEMP, M_WAITOK|M_CANFAIL);
|
|
if (buf == NULL) {
|
|
error = ENOMEM;
|
|
goto error;
|
|
}
|
|
|
|
kstat_enter(ks);
|
|
error = (*ks->ks_read)(ks);
|
|
if (error == 0) {
|
|
updated = ks->ks_updated;
|
|
|
|
/* KSTAT_F_REALLOC */
|
|
KASSERTMSG(ks->ks_datalen == klen,
|
|
"kstat doesnt support resized data yet");
|
|
|
|
error = (*ks->ks_copy)(ks, buf);
|
|
}
|
|
kstat_leave(ks);
|
|
|
|
if (error != 0)
|
|
goto error;
|
|
|
|
ulen = ksreq->ks_datalen;
|
|
ksreq->ks_datalen = klen; /* KSTAT_F_REALLOC */
|
|
ksreq->ks_updated = updated;
|
|
break;
|
|
default:
|
|
panic("ks %p unexpected state %u", ks, ks->ks_state);
|
|
}
|
|
|
|
ksreq->ks_version = kstat_version;
|
|
ksreq->ks_id = ks->ks_id;
|
|
|
|
if (strlcpy(ksreq->ks_provider, ks->ks_provider,
|
|
sizeof(ksreq->ks_provider)) >= sizeof(ksreq->ks_provider))
|
|
panic("kstat %p provider string has grown", ks);
|
|
ksreq->ks_instance = ks->ks_instance;
|
|
if (strlcpy(ksreq->ks_name, ks->ks_name,
|
|
sizeof(ksreq->ks_name)) >= sizeof(ksreq->ks_name))
|
|
panic("kstat %p name string has grown", ks);
|
|
ksreq->ks_unit = ks->ks_unit;
|
|
|
|
ksreq->ks_created = ks->ks_created;
|
|
ksreq->ks_type = ks->ks_type;
|
|
ksreq->ks_state = ks->ks_state;
|
|
|
|
error:
|
|
rw_exit(&kstat_lock);
|
|
|
|
if (buf != NULL) {
|
|
if (error == 0)
|
|
error = copyout(buf, ksreq->ks_data, min(klen, ulen));
|
|
|
|
free(buf, M_TEMP, klen);
|
|
}
|
|
|
|
return (error);
|
|
}
|
|
|
|
int
|
|
kstatioc_find_id(struct kstat_req *ksreq)
|
|
{
|
|
struct kstat *ks, key;
|
|
int error;
|
|
|
|
error = kstatioc_enter(ksreq);
|
|
if (error != 0)
|
|
return (error);
|
|
|
|
key.ks_id = ksreq->ks_id;
|
|
|
|
ks = RBT_FIND(kstat_id_tree, &kstat_id_tree, &key);
|
|
|
|
return (kstatioc_leave(ksreq, ks));
|
|
}
|
|
|
|
int
|
|
kstatioc_nfind_id(struct kstat_req *ksreq)
|
|
{
|
|
struct kstat *ks, key;
|
|
int error;
|
|
|
|
error = kstatioc_enter(ksreq);
|
|
if (error != 0)
|
|
return (error);
|
|
|
|
key.ks_id = ksreq->ks_id;
|
|
|
|
ks = RBT_NFIND(kstat_id_tree, &kstat_id_tree, &key);
|
|
|
|
return (kstatioc_leave(ksreq, ks));
|
|
}
|
|
|
|
int
|
|
kstatioc_find_pv(struct kstat_req *ksreq)
|
|
{
|
|
struct kstat *ks, key;
|
|
int error;
|
|
|
|
error = kstatioc_enter(ksreq);
|
|
if (error != 0)
|
|
return (error);
|
|
|
|
key.ks_provider = ksreq->ks_provider;
|
|
key.ks_instance = ksreq->ks_instance;
|
|
key.ks_name = ksreq->ks_name;
|
|
key.ks_unit = ksreq->ks_unit;
|
|
|
|
ks = RBT_FIND(kstat_pv_tree, &kstat_pv_tree, &key);
|
|
|
|
return (kstatioc_leave(ksreq, ks));
|
|
}
|
|
|
|
int
|
|
kstatioc_nfind_pv(struct kstat_req *ksreq)
|
|
{
|
|
struct kstat *ks, key;
|
|
int error;
|
|
|
|
error = kstatioc_enter(ksreq);
|
|
if (error != 0)
|
|
return (error);
|
|
|
|
key.ks_provider = ksreq->ks_provider;
|
|
key.ks_instance = ksreq->ks_instance;
|
|
key.ks_name = ksreq->ks_name;
|
|
key.ks_unit = ksreq->ks_unit;
|
|
|
|
ks = RBT_NFIND(kstat_pv_tree, &kstat_pv_tree, &key);
|
|
|
|
return (kstatioc_leave(ksreq, ks));
|
|
}
|
|
|
|
int
|
|
kstatioc_find_nm(struct kstat_req *ksreq)
|
|
{
|
|
struct kstat *ks, key;
|
|
int error;
|
|
|
|
error = kstatioc_enter(ksreq);
|
|
if (error != 0)
|
|
return (error);
|
|
|
|
key.ks_name = ksreq->ks_name;
|
|
key.ks_unit = ksreq->ks_unit;
|
|
key.ks_provider = ksreq->ks_provider;
|
|
key.ks_instance = ksreq->ks_instance;
|
|
|
|
ks = RBT_FIND(kstat_nm_tree, &kstat_nm_tree, &key);
|
|
|
|
return (kstatioc_leave(ksreq, ks));
|
|
}
|
|
|
|
int
|
|
kstatioc_nfind_nm(struct kstat_req *ksreq)
|
|
{
|
|
struct kstat *ks, key;
|
|
int error;
|
|
|
|
error = kstatioc_enter(ksreq);
|
|
if (error != 0)
|
|
return (error);
|
|
|
|
key.ks_name = ksreq->ks_name;
|
|
key.ks_unit = ksreq->ks_unit;
|
|
key.ks_provider = ksreq->ks_provider;
|
|
key.ks_instance = ksreq->ks_instance;
|
|
|
|
ks = RBT_NFIND(kstat_nm_tree, &kstat_nm_tree, &key);
|
|
|
|
return (kstatioc_leave(ksreq, ks));
|
|
}
|
|
|
|
int
|
|
kstatioctl(dev_t dev, u_long cmd, caddr_t data, int flag, struct proc *p)
|
|
{
|
|
struct kstat_req *ksreq = (struct kstat_req *)data;
|
|
int error = 0;
|
|
|
|
KERNEL_UNLOCK();
|
|
|
|
switch (cmd) {
|
|
case KSTATIOC_VERSION:
|
|
*(unsigned int *)data = kstat_version;
|
|
break;
|
|
|
|
case KSTATIOC_FIND_ID:
|
|
error = kstatioc_find_id(ksreq);
|
|
break;
|
|
case KSTATIOC_NFIND_ID:
|
|
error = kstatioc_nfind_id(ksreq);
|
|
break;
|
|
case KSTATIOC_FIND_PROVIDER:
|
|
error = kstatioc_find_pv(ksreq);
|
|
break;
|
|
case KSTATIOC_NFIND_PROVIDER:
|
|
error = kstatioc_nfind_pv(ksreq);
|
|
break;
|
|
case KSTATIOC_FIND_NAME:
|
|
error = kstatioc_find_nm(ksreq);
|
|
break;
|
|
case KSTATIOC_NFIND_NAME:
|
|
error = kstatioc_nfind_nm(ksreq);
|
|
break;
|
|
|
|
default:
|
|
error = ENOTTY;
|
|
break;
|
|
}
|
|
|
|
KERNEL_LOCK();
|
|
|
|
return (error);
|
|
}
|
|
|
|
void
|
|
kstat_init(void)
|
|
{
|
|
static int initialized = 0;
|
|
|
|
if (initialized)
|
|
return;
|
|
|
|
pool_init(&kstat_pool, sizeof(struct kstat), 0, IPL_NONE,
|
|
PR_WAITOK | PR_RWLOCK, "kstatmem", NULL);
|
|
|
|
initialized = 1;
|
|
}
|
|
|
|
int
|
|
kstat_strcheck(const char *str)
|
|
{
|
|
size_t i, l;
|
|
|
|
l = strlen(str);
|
|
if (l == 0 || l >= KSTAT_STRLEN)
|
|
return (-1);
|
|
for (i = 0; i < l; i++) {
|
|
int ch = str[i];
|
|
if (ch >= 'a' && ch <= 'z')
|
|
continue;
|
|
if (ch >= 'A' && ch <= 'Z')
|
|
continue;
|
|
if (ch >= '0' && ch <= '9')
|
|
continue;
|
|
switch (ch) {
|
|
case '-':
|
|
case '_':
|
|
case '.':
|
|
break;
|
|
default:
|
|
return (-1);
|
|
}
|
|
}
|
|
|
|
return (0);
|
|
}
|
|
|
|
struct kstat *
|
|
kstat_create(const char *provider, unsigned int instance,
|
|
const char *name, unsigned int unit,
|
|
unsigned int type, unsigned int flags)
|
|
{
|
|
struct kstat *ks, *oks;
|
|
|
|
if (kstat_strcheck(provider) == -1)
|
|
panic("invalid provider string");
|
|
if (kstat_strcheck(name) == -1)
|
|
panic("invalid name string");
|
|
|
|
kstat_init();
|
|
|
|
ks = pool_get(&kstat_pool, PR_WAITOK|PR_ZERO);
|
|
|
|
ks->ks_provider = provider;
|
|
ks->ks_instance = instance;
|
|
ks->ks_name = name;
|
|
ks->ks_unit = unit;
|
|
ks->ks_flags = flags;
|
|
ks->ks_type = type;
|
|
ks->ks_state = KSTAT_S_CREATED;
|
|
|
|
getnanouptime(&ks->ks_created);
|
|
ks->ks_updated = ks->ks_created;
|
|
|
|
ks->ks_lock = &kstat_default_lock;
|
|
ks->ks_lock_ops = &kstat_wlock_ops;
|
|
ks->ks_read = kstat_read;
|
|
ks->ks_copy = kstat_copy;
|
|
|
|
rw_enter_write(&kstat_lock);
|
|
ks->ks_id = kstat_next_id;
|
|
|
|
oks = RBT_INSERT(kstat_pv_tree, &kstat_pv_tree, ks);
|
|
if (oks == NULL) {
|
|
/* commit */
|
|
kstat_next_id++;
|
|
kstat_version++;
|
|
|
|
oks = RBT_INSERT(kstat_nm_tree, &kstat_nm_tree, ks);
|
|
if (oks != NULL)
|
|
panic("kstat name collision! (%llu)", ks->ks_id);
|
|
|
|
oks = RBT_INSERT(kstat_id_tree, &kstat_id_tree, ks);
|
|
if (oks != NULL)
|
|
panic("kstat id collision! (%llu)", ks->ks_id);
|
|
}
|
|
rw_exit_write(&kstat_lock);
|
|
|
|
if (oks != NULL) {
|
|
pool_put(&kstat_pool, ks);
|
|
return (NULL);
|
|
}
|
|
|
|
return (ks);
|
|
}
|
|
|
|
void
|
|
kstat_set_rlock(struct kstat *ks, struct rwlock *rwl)
|
|
{
|
|
KASSERT(ks->ks_state == KSTAT_S_CREATED);
|
|
|
|
ks->ks_lock = rwl;
|
|
ks->ks_lock_ops = &kstat_rlock_ops;
|
|
}
|
|
|
|
void
|
|
kstat_set_wlock(struct kstat *ks, struct rwlock *rwl)
|
|
{
|
|
KASSERT(ks->ks_state == KSTAT_S_CREATED);
|
|
|
|
ks->ks_lock = rwl;
|
|
ks->ks_lock_ops = &kstat_wlock_ops;
|
|
}
|
|
|
|
void
|
|
kstat_set_mutex(struct kstat *ks, struct mutex *mtx)
|
|
{
|
|
KASSERT(ks->ks_state == KSTAT_S_CREATED);
|
|
|
|
ks->ks_lock = mtx;
|
|
ks->ks_lock_ops = &kstat_mutex_ops;
|
|
}
|
|
|
|
void
|
|
kstat_cpu_enter(void *p)
|
|
{
|
|
struct cpu_info *ci = p;
|
|
sched_peg_curproc(ci);
|
|
}
|
|
|
|
void
|
|
kstat_cpu_leave(void *p)
|
|
{
|
|
atomic_clearbits_int(&curproc->p_flag, P_CPUPEG);
|
|
}
|
|
|
|
void
|
|
kstat_set_cpu(struct kstat *ks, struct cpu_info *ci)
|
|
{
|
|
KASSERT(ks->ks_state == KSTAT_S_CREATED);
|
|
|
|
ks->ks_lock = ci;
|
|
ks->ks_lock_ops = &kstat_cpu_ops;
|
|
}
|
|
|
|
int
|
|
kstat_read_nop(struct kstat *ks)
|
|
{
|
|
return (0);
|
|
}
|
|
|
|
void
|
|
kstat_install(struct kstat *ks)
|
|
{
|
|
if (!ISSET(ks->ks_flags, KSTAT_F_REALLOC)) {
|
|
KASSERTMSG(ks->ks_copy != NULL || ks->ks_data != NULL,
|
|
"kstat %p %s:%u:%s:%u must provide ks_copy or ks_data", ks,
|
|
ks->ks_provider, ks->ks_instance, ks->ks_name, ks->ks_unit);
|
|
KASSERT(ks->ks_datalen > 0);
|
|
}
|
|
|
|
rw_enter_write(&kstat_lock);
|
|
ks->ks_state = KSTAT_S_INSTALLED;
|
|
rw_exit_write(&kstat_lock);
|
|
}
|
|
|
|
void
|
|
kstat_remove(struct kstat *ks)
|
|
{
|
|
rw_enter_write(&kstat_lock);
|
|
KASSERTMSG(ks->ks_state == KSTAT_S_INSTALLED,
|
|
"kstat %p %s:%u:%s:%u is not installed", ks,
|
|
ks->ks_provider, ks->ks_instance, ks->ks_name, ks->ks_unit);
|
|
|
|
ks->ks_state = KSTAT_S_CREATED;
|
|
rw_exit_write(&kstat_lock);
|
|
}
|
|
|
|
void
|
|
kstat_destroy(struct kstat *ks)
|
|
{
|
|
rw_enter_write(&kstat_lock);
|
|
RBT_REMOVE(kstat_id_tree, &kstat_id_tree, ks);
|
|
RBT_REMOVE(kstat_pv_tree, &kstat_pv_tree, ks);
|
|
RBT_REMOVE(kstat_nm_tree, &kstat_nm_tree, ks);
|
|
kstat_version++;
|
|
rw_exit_write(&kstat_lock);
|
|
|
|
pool_put(&kstat_pool, ks);
|
|
}
|
|
|
|
int
|
|
kstat_read(struct kstat *ks)
|
|
{
|
|
getnanouptime(&ks->ks_updated);
|
|
return (0);
|
|
}
|
|
|
|
int
|
|
kstat_copy(struct kstat *ks, void *buf)
|
|
{
|
|
memcpy(buf, ks->ks_data, ks->ks_datalen);
|
|
return (0);
|
|
}
|
|
|
|
RBT_GENERATE(kstat_id_tree, kstat, ks_id_entry, kstat_id_cmp);
|
|
RBT_GENERATE(kstat_pv_tree, kstat, ks_pv_entry, kstat_pv_cmp);
|
|
RBT_GENERATE(kstat_nm_tree, kstat, ks_nm_entry, kstat_nm_cmp);
|
|
|
|
void
|
|
kstat_kv_init(struct kstat_kv *kv, const char *name, enum kstat_kv_type type)
|
|
{
|
|
memset(kv, 0, sizeof(*kv));
|
|
strlcpy(kv->kv_key, name, sizeof(kv->kv_key)); /* XXX truncated? */
|
|
kv->kv_type = type;
|
|
kv->kv_unit = KSTAT_KV_U_NONE;
|
|
}
|
|
|
|
void
|
|
kstat_kv_unit_init(struct kstat_kv *kv, const char *name,
|
|
enum kstat_kv_type type, enum kstat_kv_unit unit)
|
|
{
|
|
switch (type) {
|
|
case KSTAT_KV_T_COUNTER64:
|
|
case KSTAT_KV_T_COUNTER32:
|
|
case KSTAT_KV_T_UINT64:
|
|
case KSTAT_KV_T_INT64:
|
|
case KSTAT_KV_T_UINT32:
|
|
case KSTAT_KV_T_INT32:
|
|
break;
|
|
default:
|
|
panic("kv unit init %s: unit for non-integer type", name);
|
|
}
|
|
|
|
memset(kv, 0, sizeof(*kv));
|
|
strlcpy(kv->kv_key, name, sizeof(kv->kv_key)); /* XXX truncated? */
|
|
kv->kv_type = type;
|
|
kv->kv_unit = unit;
|
|
}
|