#endif
/* use this to force every realloc to change the pointer, to stress test
code that might not cope */
#define ALWAYS_REALLOC 0
#define MAX_TALLOC_SIZE 0x10000000
#define TALLOC_MAGIC_BASE 0xe814ec70
#define TALLOC_MAGIC ( \
TALLOC_MAGIC_BASE + \
(TALLOC_VERSION_MAJOR << 12) + \
(TALLOC_VERSION_MINOR << 4) \
)
#define TALLOC_FLAG_FREE 0x01
#define TALLOC_FLAG_LOOP 0x02
#define TALLOC_FLAG_POOL 0x04 /* This is a talloc pool */
#define TALLOC_FLAG_POOLMEM 0x08 /* This is allocated in a pool */
#define TALLOC_MAGIC_REFERENCE ((const char *)1)
/* by default we abort when given a bad pointer (such as when talloc_free() is called
on a pointer that came from malloc() */
#ifndef TALLOC_ABORT
#define TALLOC_ABORT(reason) abort()
#endif
#ifndef discard_const_p
#if defined(__intptr_t_defined) || defined(HAVE_INTPTR_T)
# define discard_const_p(type, ptr) ((type *)((intptr_t)(ptr)))
#else
# define discard_const_p(type, ptr) ((type *)(ptr))
#endif
#endif
/* these macros gain us a few percent of speed on gcc */
#if (__GNUC__ >= 3)
/* the strange !! is to ensure that __builtin_expect() takes either 0 or 1
as its first argument */
#ifndef likely
#define likely(x) __builtin_expect(!!(x), 1)
#endif
#ifndef unlikely
#define unlikely(x) __builtin_expect(!!(x), 0)
#endif
#else
#ifndef likely
#define likely(x) (x)
#endif
#ifndef unlikely
#define unlikely(x) (x)
#endif
#endif
/* this null_context is only used if talloc_enable_leak_report() or
talloc_enable_leak_report_full() is called, otherwise it remains
NULL
*/
static void *null_context;
static void *autofree_context;
/* used to enable fill of memory on free, which can be useful for
* catching use after free errors when valgrind is too slow
*/
static struct {
bool initialised;
bool enabled;
uint8_t fill_value;
} talloc_fill;
#define TALLOC_FILL_ENV "TALLOC_FREE_FILL"
/*
* do not wipe the header, to allow the
* double-free logic to still work
*/
#define TC_INVALIDATE_FULL_FILL_CHUNK(_tc) do { \
if (unlikely(talloc_fill.enabled)) { \
size_t _flen = (_tc)->size; \
char *_fptr = (char *)TC_PTR_FROM_CHUNK(_tc); \
memset(_fptr, talloc_fill.fill_value, _flen); \
} \
} while (0)
#if defined(DEVELOPER) && defined(VALGRIND_MAKE_MEM_NOACCESS)
/* Mark the whole chunk as not accessable */
#define TC_INVALIDATE_FULL_VALGRIND_CHUNK(_tc) do { \
size_t _flen = TC_HDR_SIZE + (_tc)->size; \
char *_fptr = (char *)(_tc); \
VALGRIND_MAKE_MEM_NOACCESS(_fptr, _flen); \
} while(0)
#else
#define TC_INVALIDATE_FULL_VALGRIND_CHUNK(_tc) do { } while (0)
#endif
#define TC_INVALIDATE_FULL_CHUNK(_tc) do { \
TC_INVALIDATE_FULL_FILL_CHUNK(_tc); \
TC_INVALIDATE_FULL_VALGRIND_CHUNK(_tc); \
} while (0)
#define TC_INVALIDATE_SHRINK_FILL_CHUNK(_tc, _new_size) do { \
if (unlikely(talloc_fill.enabled)) { \
size_t _flen = (_tc)->size - (_new_size); \
char *_fptr = (char *)TC_PTR_FROM_CHUNK(_tc); \
_fptr += (_new_size); \
memset(_fptr, talloc_fill.fill_value, _flen); \
} \
} while (0)
#if defined(DEVELOPER) && defined(VALGRIND_MAKE_MEM_NOACCESS)
/* Mark the unused bytes not accessable */
#define TC_INVALIDATE_SHRINK_VALGRIND_CHUNK(_tc, _new_size) do { \
size_t _flen = (_tc)->size - (_new_size); \
char *_fptr = (char *)TC_PTR_FROM_CHUNK(_tc); \
_fptr += (_new_size); \
VALGRIND_MAKE_MEM_NOACCESS(_fptr, _flen); \
} while (0)
#else
#define TC_INVALIDATE_SHRINK_VALGRIND_CHUNK(_tc, _new_size) do { } while (0)
#endif
#define TC_INVALIDATE_SHRINK_CHUNK(_tc, _new_size) do { \
TC_INVALIDATE_SHRINK_FILL_CHUNK(_tc, _new_size); \
TC_INVALIDATE_SHRINK_VALGRIND_CHUNK(_tc, _new_size); \
} while (0)
#define TC_UNDEFINE_SHRINK_FILL_CHUNK(_tc, _new_size) do { \
if (unlikely(talloc_fill.enabled)) { \
size_t _flen = (_tc)->size - (_new_size); \
char *_fptr = (char *)TC_PTR_FROM_CHUNK(_tc); \
_fptr += (_new_size); \
memset(_fptr, talloc_fill.fill_value, _flen); \
} \
} while (0)
#if defined(DEVELOPER) && defined(VALGRIND_MAKE_MEM_UNDEFINED)
/* Mark the unused bytes as undefined */
#define TC_UNDEFINE_SHRINK_VALGRIND_CHUNK(_tc, _new_size) do { \
size_t _flen = (_tc)->size - (_new_size); \
char *_fptr = (char *)TC_PTR_FROM_CHUNK(_tc); \
_fptr += (_new_size); \
VALGRIND_MAKE_MEM_UNDEFINED(_fptr, _flen); \
} while (0)
#else
#define TC_UNDEFINE_SHRINK_VALGRIND_CHUNK(_tc, _new_size) do { } while (0)
#endif
#define TC_UNDEFINE_SHRINK_CHUNK(_tc, _new_size) do { \
TC_UNDEFINE_SHRINK_FILL_CHUNK(_tc, _new_size); \
TC_UNDEFINE_SHRINK_VALGRIND_CHUNK(_tc, _new_size); \
} while (0)
#if defined(DEVELOPER) && defined(VALGRIND_MAKE_MEM_UNDEFINED)
/* Mark the new bytes as undefined */
#define TC_UNDEFINE_GROW_VALGRIND_CHUNK(_tc, _new_size) do { \
size_t _old_used = TC_HDR_SIZE + (_tc)->size; \
size_t _new_used = TC_HDR_SIZE + (_new_size); \
size_t _flen = _new_used - _old_used; \
char *_fptr = _old_used + (char *)(_tc); \
VALGRIND_MAKE_MEM_UNDEFINED(_fptr, _flen); \
} while (0)
#else
#define TC_UNDEFINE_GROW_VALGRIND_CHUNK(_tc, _new_size) do { } while (0)
#endif
#define TC_UNDEFINE_GROW_CHUNK(_tc, _new_size) do { \
TC_UNDEFINE_GROW_VALGRIND_CHUNK(_tc, _new_size); \
} while (0)
struct talloc_reference_handle {
struct talloc_reference_handle *next, *prev;
void *ptr;
const char *location;
};
struct talloc_memlimit {
struct talloc_chunk *parent;
struct talloc_memlimit *upper;
size_t max_size;
size_t cur_size;
};
static bool talloc_memlimit_check(struct talloc_memlimit *limit, size_t size);
static bool talloc_memlimit_update(struct talloc_memlimit *limit,
size_t old_size, size_t new_size);
typedef int (*talloc_destructor_t)(void *);
struct talloc_chunk {
struct talloc_chunk *next, *prev;
struct talloc_chunk *parent, *child;
struct talloc_reference_handle *refs;
talloc_destructor_t destructor;
const char *name;
size_t size;
unsigned flags;
/*
* limit semantics:
* if 'limit' is set it means all *new* children of the context will
* be limited to a total aggregate size ox max_size for memory
* allocations.
* cur_size is used to keep track of the current use
*/
struct talloc_memlimit *limit;
/*
* "pool" has dual use:
*
* For the talloc pool itself (i.e. TALLOC_FLAG_POOL is set), "pool"
* marks the end of the currently allocated area.
*
* For members of the pool (i.e. TALLOC_FLAG_POOLMEM is set), "pool"
* is a pointer to the struct talloc_chunk of the pool that it was
* allocated from. This way children can quickly find the pool to chew
* from.
*/
void *pool;
};
/* 16 byte alignment seems to keep everyone happy */
#define TC_ALIGN16(s) (((s)+15)&~15)
#define TC_HDR_SIZE TC_ALIGN16(sizeof(struct talloc_chunk))
#define TC_PTR_FROM_CHUNK(tc) ((void *)(TC_HDR_SIZE + (char*)tc))
_PUBLIC_ int talloc_version_major(void)
{
return TALLOC_VERSION_MAJOR;
}
_PUBLIC_ int talloc_version_minor(void)
{
return TALLOC_VERSION_MINOR;
}
static void (*talloc_log_fn)(const char *message);
_PUBLIC_ void talloc_set_log_fn(void (*log_fn)(const char *message))
{
talloc_log_fn = log_fn;
}
static void talloc_log(const char *fmt, ...) PRINTF_ATTRIBUTE(1,2);
static void talloc_log(const char *fmt, ...)
{
va_list ap;
char *message;
if (!talloc_log_fn) {
return;
}
va_start(ap, fmt);
message = talloc_vasprintf(NULL, fmt, ap);
va_end(ap);
talloc_log_fn(message);
talloc_free(message);
}
static void talloc_log_stderr(const char *message)
{
fprintf(stderr, "%s", message);
}
_PUBLIC_ void talloc_set_log_stderr(void)
{
talloc_set_log_fn(talloc_log_stderr);
}
static void (*talloc_abort_fn)(const char *reason);
_PUBLIC_ void talloc_set_abort_fn(void (*abort_fn)(const char *reason))
{
talloc_abort_fn = abort_fn;
}
static void talloc_abort(const char *reason)
{
talloc_log("%s\n", reason);
if (!talloc_abort_fn) {
TALLOC_ABORT(reason);
}
talloc_abort_fn(reason);
}
static void talloc_abort_magic(unsigned magic)
{
unsigned striped = magic - TALLOC_MAGIC_BASE;
unsigned major = (striped & 0xFFFFF000) >> 12;
unsigned minor = (striped & 0x00000FF0) >> 4;
talloc_log("Bad talloc magic[0x%08X/%u/%u] expected[0x%08X/%u/%u]\n",
magic, major, minor,
TALLOC_MAGIC, TALLOC_VERSION_MAJOR, TALLOC_VERSION_MINOR);
talloc_abort("Bad talloc magic value - wrong talloc version used/mixed");
}
static void talloc_abort_access_after_free(void)
{
talloc_abort("Bad talloc magic value - access after free");
}
static void talloc_abort_unknown_value(void)
{
talloc_abort("Bad talloc magic value - unknown value");
}
/* panic if we get a bad magic value */
static inline struct talloc_chunk *talloc_chunk_from_ptr(const void *ptr)
{
const char *pp = (const char *)ptr;
struct talloc_chunk *tc = discard_const_p(struct talloc_chunk, pp - TC_HDR_SIZE);
if (unlikely((tc->flags & (TALLOC_FLAG_FREE | ~0xF)) != TALLOC_MAGIC)) {
if ((tc->flags & (~0xFFF)) == TALLOC_MAGIC_BASE) {
talloc_abort_magic(tc->flags & (~0xF));
return NULL;
}
if (tc->flags & TALLOC_FLAG_FREE) {
talloc_log("talloc: access after free error - first free may be at %s\n", tc->name);
talloc_abort_access_after_free();
return NULL;
} else {
talloc_abort_unknown_value();
return NULL;
}
}
return tc;
}
/* hook into the front of the list */
#define _TLIST_ADD(list, p) \
do { \
if (!(list)) { \
(list) = (p); \
(p)->next = (p)->prev = NULL; \
} else { \
(list)->prev = (p); \
(p)->next = (list); \
(p)->prev = NULL; \
(list) = (p); \
}\
} while (0)
/* remove an element from a list - element doesn't have to be in list. */
#define _TLIST_REMOVE(list, p) \
do { \
if ((p) == (list)) { \
(list) = (p)->next; \
if (list) (list)->prev = NULL; \
} else { \
if ((p)->prev) (p)->prev->next = (p)->next; \
if ((p)->next) (p)->next->prev = (p)->prev; \
} \
if ((p) && ((p) != (list))) (p)->next = (p)->prev = NULL; \
} while (0)
/*
return the parent chunk of a pointer
*/
static inline struct talloc_chunk *talloc_parent_chunk(const void *ptr)
{
struct talloc_chunk *tc;
if (unlikely(ptr == NULL)) {
return NULL;
}
tc = talloc_chunk_from_ptr(ptr);
while (tc->prev) tc=tc->prev;
return tc->parent;
}
_PUBLIC_ void *talloc_parent(const void *ptr)
{
struct talloc_chunk *tc = talloc_parent_chunk(ptr);
return tc? TC_PTR_FROM_CHUNK(tc) : NULL;
}
/*
find parents name
*/
_PUBLIC_ const char *talloc_parent_name(const void *ptr)
{
struct talloc_chunk *tc = talloc_parent_chunk(ptr);
return tc? tc->name : NULL;
}
/*
A pool carries an in-pool object count count in the first 16 bytes.
bytes. This is done to support talloc_steal() to a parent outside of the
pool. The count includes the pool itself, so a talloc_free() on a pool will
only destroy the pool if the count has dropped to zero. A talloc_free() of a
pool member will reduce the count, and eventually also call free(3) on the
pool memory.
The object count is not put into "struct talloc_chunk" because it is only
relevant for talloc pools and the alignment to 16 bytes would increase the
memory footprint of each talloc chunk by those 16 bytes.
*/
union talloc_pool_chunk {
/* This lets object_count nestle into 16-byte padding of talloc_chunk,
* on 32-bit platforms. */
struct tc_pool_hdr {
struct talloc_chunk c;
unsigned int object_count;
} hdr;
/* This makes it always 16 byte aligned. */
char pad[TC_ALIGN16(sizeof(struct tc_pool_hdr))];
};
static void *tc_pool_end(union talloc_pool_chunk *pool_tc)
{
return (char *)pool_tc + TC_HDR_SIZE + pool_tc->hdr.c.size;
}
static size_t tc_pool_space_left(union talloc_pool_chunk *pool_tc)
{
return (char *)tc_pool_end(pool_tc) - (char *)pool_tc->hdr.c.pool;
}
static void *tc_pool_first_chunk(union talloc_pool_chunk *pool_tc)
{
return pool_tc + 1;
}
/* If tc is inside a pool, this gives the next neighbour. */
static void *tc_next_chunk(struct talloc_chunk *tc)
{
return (char *)tc + TC_ALIGN16(TC_HDR_SIZE + tc->size);
}
/* Mark the whole remaining pool as not accessable */
static void tc_invalidate_pool(union talloc_pool_chunk *pool_tc)
{
size_t flen = tc_pool_space_left(pool_tc);
if (unlikely(talloc_fill.enabled)) {
memset(pool_tc->hdr.c.pool, talloc_fill.fill_value, flen);
}
#if defined(DEVELOPER) && defined(VALGRIND_MAKE_MEM_NOACCESS)
VALGRIND_MAKE_MEM_NOACCESS(pool_tc->hdr.c.pool, flen);
#endif
}
/*
Allocate from a pool
*/
static struct talloc_chunk *talloc_alloc_pool(struct talloc_chunk *parent,
size_t size)
{
union talloc_pool_chunk *pool_ctx = NULL;
size_t space_left;
struct talloc_chunk *result;
size_t chunk_size;
if (parent == NULL) {
return NULL;
}
if (parent->flags & TALLOC_FLAG_POOL) {
pool_ctx = (union talloc_pool_chunk *)parent;
}
else if (parent->flags & TALLOC_FLAG_POOLMEM) {
pool_ctx = (union talloc_pool_chunk *)parent->pool;
}
if (pool_ctx == NULL) {
return NULL;
}
space_left = tc_pool_space_left(pool_ctx);
/*
* Align size to 16 bytes
*/
chunk_size = TC_ALIGN16(size);
if (space_left < chunk_size) {
return NULL;
}
result = (struct talloc_chunk *)pool_ctx->hdr.c.pool;
#if defined(DEVELOPER) && defined(VALGRIND_MAKE_MEM_UNDEFINED)
VALGRIND_MAKE_MEM_UNDEFINED(result, size);
#endif
pool_ctx->hdr.c.pool = (void *)((char *)result + chunk_size);
result->flags = TALLOC_MAGIC | TALLOC_FLAG_POOLMEM;
result->pool = pool_ctx;
pool_ctx->hdr.object_count++;
return result;
}
/*
Allocate a bit of memory as a child of an existing pointer
*/
static inline void *__talloc(const void *context, size_t size)
{
struct talloc_chunk *tc = NULL;
struct talloc_memlimit *limit = NULL;
if (unlikely(context == NULL)) {
context = null_context;
}
if (unlikely(size >= MAX_TALLOC_SIZE)) {
return NULL;
}
if (context != NULL) {
struct talloc_chunk *ptc = talloc_chunk_from_ptr(context);
if (ptc->limit != NULL) {
limit = ptc->limit;
}
if (!talloc_memlimit_check(limit, (TC_HDR_SIZE+size))) {
errno = ENOMEM;
return NULL;
}
tc = talloc_alloc_pool(ptc, TC_HDR_SIZE+size);
}
if (tc == NULL) {
tc = (struct talloc_chunk *)malloc(TC_HDR_SIZE+size);
if (unlikely(tc == NULL)) return NULL;
tc->flags = TALLOC_MAGIC;
tc->pool = NULL;
}
if (limit != NULL) {
struct talloc_memlimit *l;
for (l = limit; l != NULL; l = l->upper) {
l->cur_size += TC_HDR_SIZE+size;
}
}
tc->limit = limit;
tc->size = size;
tc->destructor = NULL;
tc->child = NULL;
tc->name = NULL;
tc->refs = NULL;
if (likely(context)) {
struct talloc_chunk *parent = talloc_chunk_from_ptr(context);
if (parent->child) {
parent->child->parent = NULL;
tc->next = parent->child;
tc->next->prev = tc;
} else {
tc->next = NULL;
}
tc->parent = parent;
tc->prev = NULL;
parent->child = tc;
} else {
tc->next = tc->prev = tc->parent = NULL;
}
return TC_PTR_FROM_CHUNK(tc);
}
/*
* Create a talloc pool
*/
_PUBLIC_ void *talloc_pool(const void *context, size_t size)
{
union talloc_pool_chunk *pool_tc;
void *result = __talloc(context, sizeof(*pool_tc) - TC_HDR_SIZE + size);
if (unlikely(result == NULL)) {
return NULL;
}
pool_tc = (union talloc_pool_chunk *)talloc_chunk_from_ptr(result);
if (unlikely(pool_tc->hdr.c.flags & TALLOC_FLAG_POOLMEM)) {
/* We don't handle this correctly, so fail. */
talloc_log("talloc: cannot allocate pool off another pool %s\n",
talloc_get_name(context));
talloc_free(result);
return NULL;
}
pool_tc->hdr.c.flags |= TALLOC_FLAG_POOL;
pool_tc->hdr.c.pool = tc_pool_first_chunk(pool_tc);
pool_tc->hdr.object_count = 1;
tc_invalidate_pool(pool_tc);
return result;
}
/*
setup a destructor to be called on free of a pointer
the destructor should return 0 on success, or -1 on failure.
if the destructor fails then the free is failed, and the memory can
be continued to be used
*/
_PUBLIC_ void _talloc_set_destructor(const void *ptr, int (*destructor)(void *))
{
struct talloc_chunk *tc = talloc_chunk_from_ptr(ptr);
tc->destructor = destructor;
}
/*
increase the reference count on a piece of memory.
*/
_PUBLIC_ int talloc_increase_ref_count(const void *ptr)
{
if (unlikely(!talloc_reference(null_context, ptr))) {
return -1;
}
return 0;
}
/*
helper for talloc_reference()
this is referenced by a function pointer and should not be inline
*/
static int talloc_reference_destructor(struct talloc_reference_handle *handle)
{
struct talloc_chunk *ptr_tc = talloc_chunk_from_ptr(handle->ptr);
_TLIST_REMOVE(ptr_tc->refs, handle);
return 0;
}
/*
more efficient way to add a name to a pointer - the name must point to a
true string constant
*/
static inline void _talloc_set_name_const(const void *ptr, const char *name)
{
struct talloc_chunk *tc = talloc_chunk_from_ptr(ptr);
tc->name = name;
}
/*
internal talloc_named_const()
*/
static inline void *_talloc_named_const(const void *context, size_t size, const char *name)
{
void *ptr;
ptr = __talloc(context, size);
if (unlikely(ptr == NULL)) {
return NULL;
}
_talloc_set_name_const(ptr, name);
return ptr;
}
/*
make a secondary reference to a pointer, hanging off the given context.
the pointer remains valid until both the original caller and this given
context are freed.
the major use for this is when two different structures need to reference the
same underlying data, and you want to be able to free the two instances separately,
and in either order
*/
_PUBLIC_ void *_talloc_reference_loc(const void *context, const void *ptr, const char *location)
{
struct talloc_chunk *tc;
struct talloc_reference_handle *handle;
if (unlikely(ptr == NULL)) return NULL;
tc = talloc_chunk_from_ptr(ptr);
handle = (struct talloc_reference_handle *)_talloc_named_const(context,
sizeof(struct talloc_reference_handle),
TALLOC_MAGIC_REFERENCE);
if (unlikely(handle == NULL)) return NULL;
/* note that we hang the destructor off the handle, not the
main context as that allows the caller to still setup their
own destructor on the context if they want to */
talloc_set_destructor(handle, talloc_reference_destructor);
handle->ptr = discard_const_p(void, ptr);
handle->location = location;
_TLIST_ADD(tc->refs, handle);
return handle->ptr;
}
static void *_talloc_steal_internal(const void *new_ctx, const void *ptr);
static inline void _talloc_free_poolmem(struct talloc_chunk *tc,
const char *location)
{
union talloc_pool_chunk *pool;
void *next_tc;
pool = (union talloc_pool_chunk *)tc->pool;
next_tc = tc_next_chunk(tc);
tc->flags |= TALLOC_FLAG_FREE;
/* we mark the freed memory with where we called the free
* from. This means on a double free error we can report where
* the first free came from
*/
tc->name = location;
TC_INVALIDATE_FULL_CHUNK(tc);
if (unlikely(pool->hdr.object_count == 0)) {
talloc_abort("Pool object count zero!");
return;
}
pool->hdr.object_count--;
if (unlikely(pool->hdr.object_count == 1
&& !(pool->hdr.c.flags & TALLOC_FLAG_FREE))) {
/*
* if there is just one object left in the pool
* and pool->flags does not have TALLOC_FLAG_FREE,
* it means this is the pool itself and
* the rest is available for new objects
* again.
*/
pool->hdr.c.pool = tc_pool_first_chunk(pool);
tc_invalidate_pool(pool);
return;
}
if (unlikely(pool->hdr.object_count == 0)) {
/*
* we mark the freed memory with where we called the free
* from. This means on a double free error we can report where
* the first free came from
*/
pool->hdr.c.name = location;
TC_INVALIDATE_FULL_CHUNK(&pool->hdr.c);
free(pool);
return;
}
if (pool->hdr.c.pool == next_tc) {
/*
* if pool->pool still points to end of
* 'tc' (which is stored in the 'next_tc' variable),
* we can reclaim the memory of 'tc'.
*/
pool->hdr.c.pool = tc;
return;
}
/*
* Do nothing. The memory is just "wasted", waiting for the pool
* itself to be freed.
*/
}
static inline void _talloc_free_children_internal(struct talloc_chunk *tc,
void *ptr,
const char *location);
/*
internal talloc_free call
*/
static inline int _talloc_free_internal(void *ptr, const char *location)
{
struct talloc_chunk *tc;
if (unlikely(ptr == NULL)) {
return -1;
}
/* possibly initialised the talloc fill value */
if (unlikely(!talloc_fill.initialised)) {
const char *fill = getenv(TALLOC_FILL_ENV);
if (fill != NULL) {
talloc_fill.enabled = true;
talloc_fill.fill_value = strtoul(fill, NULL, 0);
}
talloc_fill.initialised = true;
}
tc = talloc_chunk_from_ptr(ptr);
if (unlikely(tc->refs)) {
int is_child;
/* check if this is a reference from a child or
* grandchild back to it's parent or grandparent
*
* in that case we need to remove the reference and
* call another instance of talloc_free() on the current
* pointer.
*/
is_child = talloc_is_parent(tc->refs, ptr);
_talloc_free_internal(tc->refs, location);
if (is_child) {
return _talloc_free_internal(ptr, location);
}
return -1;
}
if (unlikely(tc->flags & TALLOC_FLAG_LOOP)) {
/* we have a free loop - stop looping */
return 0;
}
if (unlikely(tc->destructor)) {
talloc_destructor_t d = tc->destructor;
if (d == (talloc_destructor_t)-1) {
return -1;
}
tc->destructor = (talloc_destructor_t)-1;
if (d(ptr) == -1) {
tc->destructor = d;
return -1;
}
tc->destructor = NULL;
}
if (tc->parent) {
_TLIST_REMOVE(tc->parent->child, tc);
if (tc->parent->child) {
tc->parent->child->parent = tc->parent;
}
} else {
if (tc->prev) tc->prev->next = tc->next;
if (tc->next) tc->next->prev = tc->prev;
tc->prev = tc->next = NULL;
}
tc->flags |= TALLOC_FLAG_LOOP;
_talloc_free_children_internal(tc, ptr, location);
tc->flags |= TALLOC_FLAG_FREE;
/*
* If we are part of a memory limited context hierarchy
* we need to subtract the memory used from the counters
*/
if (tc->limit) {
struct talloc_memlimit *l;
for (l = tc->limit; l != NULL; l = l->upper) {
if (l->cur_size >= tc->size+TC_HDR_SIZE) {
l->cur_size -= tc->size+TC_HDR_SIZE;
} else {
talloc_abort("cur_size memlimit counter not correct!");
return 0;
}
}
if (tc->limit->parent == tc) {
free(tc->limit);
}
tc->limit = NULL;
}
/* we mark the freed memory with where we called the free
* from. This means on a double free error we can report where
* the first free came from
*/
tc->name = location;
if (tc->flags & TALLOC_FLAG_POOL) {
union talloc_pool_chunk *pool = (union talloc_pool_chunk *)tc;
if (unlikely(pool->hdr.object_count == 0)) {
talloc_abort("Pool object count zero!");
return 0;
}
pool->hdr.object_count--;
if (likely(pool->hdr.object_count != 0)) {
return 0;
}
TC_INVALIDATE_FULL_CHUNK(tc);
free(tc);
return 0;
}
if (tc->flags & TALLOC_FLAG_POOLMEM) {
_talloc_free_poolmem(tc, location);
return 0;
}
TC_INVALIDATE_FULL_CHUNK(tc);
free(tc);
return 0;
}
static size_t _talloc_total_limit_size(const void *ptr,
struct talloc_memlimit *old_limit,
struct talloc_memlimit *new_limit);
/*
move a lump of memory from one talloc context to another return the
ptr on success, or NULL if it could not be transferred.
passing NULL as ptr will always return NULL with no side effects.
*/
static void *_talloc_steal_internal(const void *new_ctx, const void *ptr)
{
struct talloc_chunk *tc, *new_tc;
size_t ctx_size = 0;
if (unlikely(!ptr)) {
return NULL;
}
if (unlikely(new_ctx == NULL)) {
new_ctx = null_context;
}
tc = talloc_chunk_from_ptr(ptr);
if (tc->limit != NULL) {
ctx_size = _talloc_total_limit_size(ptr, NULL, NULL);
if (!talloc_memlimit_update(tc->limit->upper, ctx_size, 0)) {
talloc_abort("cur_size memlimit counter not correct!");
errno = EINVAL;
return NULL;
}
if (tc->limit->parent == tc) {
tc->limit->upper = NULL;
} else {
tc->limit = NULL;
}
}
if (unlikely(new_ctx == NULL)) {
if (tc->parent) {
_TLIST_REMOVE(tc->parent->child, tc);
if (tc->parent->child) {
tc->parent->child->parent = tc->parent;
}
} else {
if (tc->prev) tc->prev->next = tc->next;
if (tc->next) tc->next->prev = tc->prev;
}
tc->parent = tc->next = tc->prev = NULL;
return discard_const_p(void, ptr);
}
new_tc = talloc_chunk_from_ptr(new_ctx);
if (unlikely(tc == new_tc || tc->parent == new_tc)) {
return discard_const_p(void, ptr);
}
if (tc->parent) {
_TLIST_REMOVE(tc->parent->child, tc);
if (tc->parent->child) {
tc->parent->child->parent = tc->parent;
}
} else {
if (tc->prev) tc->prev->next = tc->next;
if (tc->next) tc->next->prev = tc->prev;
tc->prev = tc->next = NULL;
}
tc->parent = new_tc;
if (new_tc->child) new_tc->child->parent = NULL;
_TLIST_ADD(new_tc->child, tc);
if (tc->limit || new_tc->limit) {
ctx_size = _talloc_total_limit_size(ptr, tc->limit,
new_tc->limit);
}
if (new_tc->limit) {
struct talloc_memlimit *l;
for (l = new_tc->limit; l != NULL; l = l->upper) {
l->cur_size += ctx_size;
}
}
return discard_const_p(void, ptr);
}
/*
move a lump of memory from one talloc context to another return the
ptr on success, or NULL if it could not be transferred.
passing NULL as ptr will always return NULL with no side effects.
*/
_PUBLIC_ void *_talloc_steal_loc(const void *new_ctx, const void *ptr, const char *location)
{
struct talloc_chunk *tc;
if (unlikely(ptr == NULL)) {
return NULL;
}
tc = talloc_chunk_from_ptr(ptr);
if (unlikely(tc->refs != NULL) && talloc_parent(ptr) != new_ctx) {
struct talloc_reference_handle *h;
talloc_log("WARNING: talloc_steal with references at %s\n",
location);
for (h=tc->refs; h; h=h->next) {
talloc_log("\treference at %s\n",
h->location);
}
}
#if 0
/* this test is probably too expensive to have on in the
normal build, but it useful for debugging */
if (talloc_is_parent(new_ctx, ptr)) {
talloc_log("WARNING: stealing into talloc child at %s\n", location);
}
#endif
return _talloc_steal_internal(new_ctx, ptr);
}
/*
this is like a talloc_steal(), but you must supply the old
parent. This resolves the ambiguity in a talloc_steal() which is
called on a context that has more than one parent (via references)
The old parent can be either a reference or a parent
*/
_PUBLIC_ void *talloc_reparent(const void *old_parent, const void *new_parent, const void *ptr)
{
struct talloc_chunk *tc;
struct talloc_reference_handle *h;
if (unlikely(ptr == NULL)) {
return NULL;
}
if (old_parent == talloc_parent(ptr)) {
return _talloc_steal_internal(new_parent, ptr);
}
tc = talloc_chunk_from_ptr(ptr);
for (h=tc->refs;h;h=h->next) {
if (talloc_parent(h) == old_parent) {
if (_talloc_steal_internal(new_parent, h) != h) {
return NULL;
}
return discard_const_p(void, ptr);
}
}
/* it wasn't a parent */
return NULL;
}
/*
remove a secondary reference to a pointer. This undo's what
talloc_reference() has done. The context and pointer arguments
must match those given to a talloc_reference()
*/
static inline int talloc_unreference(const void *context, const void *ptr)
{
struct talloc_chunk *tc = talloc_chunk_from_ptr(ptr);
struct talloc_reference_handle *h;
if (unlikely(context == NULL)) {
context = null_context;
}
for (h=tc->refs;h;h=h->next) {
struct talloc_chunk *p = talloc_parent_chunk(h);
if (p == NULL) {
if (context == NULL) break;
} else if (TC_PTR_FROM_CHUNK(p) == context) {
break;
}
}
if (h == NULL) {
return -1;
}
return _talloc_free_internal(h, __location__);
}
/*
remove a specific parent context from a pointer. This is a more
controlled variant of talloc_free()
*/
_PUBLIC_ int talloc_unlink(const void *context, void *ptr)
{
struct talloc_chunk *tc_p, *new_p, *tc_c;
void *new_parent;
if (ptr == NULL) {
return -1;
}
if (context == NULL) {
context = null_context;
}
if (talloc_unreference(context, ptr) == 0) {
return 0;
}
if (context != NULL) {
tc_c = talloc_chunk_from_ptr(context);
} else {
tc_c = NULL;
}
if (tc_c != talloc_parent_chunk(ptr)) {
return -1;
}
tc_p = talloc_chunk_from_ptr(ptr);
if (tc_p->refs == NULL) {
return _talloc_free_internal(ptr, __location__);
}
new_p = talloc_parent_chunk(tc_p->refs);
if (new_p) {
new_parent = TC_PTR_FROM_CHUNK(new_p);
} else {
new_parent = NULL;
}
if (talloc_unreference(new_parent, ptr) != 0) {
return -1;
}
_talloc_steal_internal(new_parent, ptr);
return 0;
}
/*
add a name to an existing pointer - va_list version
*/
static inline const char *talloc_set_name_v(const void *ptr, const char *fmt, va_list ap) PRINTF_ATTRIBUTE(2,0);
static inline const char *talloc_set_name_v(const void *ptr, const char *fmt, va_list ap)
{
struct talloc_chunk *tc = talloc_chunk_from_ptr(ptr);
tc->name = talloc_vasprintf(ptr, fmt, ap);
if (likely(tc->name)) {
_talloc_set_name_const(tc->name, ".name");
}
return tc->name;
}
/*
add a name to an existing pointer
*/
_PUBLIC_ const char *talloc_set_name(const void *ptr, const char *fmt, ...)
{
const char *name;
va_list ap;
va_start(ap, fmt);
name = talloc_set_name_v(ptr, fmt, ap);
va_end(ap);
return name;
}
/*
create a named talloc pointer. Any talloc pointer can be named, and
talloc_named() operates just like talloc() except that it allows you
to name the pointer.
*/
_PUBLIC_ void *talloc_named(const void *context, size_t size, const char *fmt, ...)
{
va_list ap;
void *ptr;
const char *name;
ptr = __talloc(context, size);
if (unlikely(ptr == NULL)) return NULL;
va_start(ap, fmt);
name = talloc_set_name_v(ptr, fmt, ap);
va_end(ap);
if (unlikely(name == NULL)) {
_talloc_free_internal(ptr, __location__);
return NULL;
}
return ptr;
}
/*
return the name of a talloc ptr, or "UNNAMED"
*/
_PUBLIC_ const char *talloc_get_name(const void *ptr)
{
struct talloc_chunk *tc = talloc_chunk_from_ptr(ptr);
if (unlikely(tc->name == TALLOC_MAGIC_REFERENCE)) {
return ".reference";
}
if (likely(tc->name)) {
return tc->name;
}
return "UNNAMED";
}
/*
check if a pointer has the given name. If it does, return the pointer,
otherwise return NULL
*/
_PUBLIC_ void *talloc_check_name(const void *ptr, const char *name)
{
const char *pname;
if (unlikely(ptr == NULL)) return NULL;
pname = talloc_get_name(ptr);
if (likely(pname == name || strcmp(pname, name) == 0)) {
return discard_const_p(void, ptr);
}
return NULL;
}
static void talloc_abort_type_mismatch(const char *location,
const char *name,
const char *expected)
{
const char *reason;
reason = talloc_asprintf(NULL,
"%s: Type mismatch: name[%s] expected[%s]",
location,
name?name:"NULL",
expected);
if (!reason) {
reason = "Type mismatch";
}
talloc_abort(reason);
}
_PUBLIC_ void *_talloc_get_type_abort(const void *ptr, const char *name, const char *location)
{
const char *pname;
if (unlikely(ptr == NULL)) {
talloc_abort_type_mismatch(location, NULL, name);
return NULL;
}
pname = talloc_get_name(ptr);
if (likely(pname == name || strcmp(pname, name) == 0)) {
return discard_const_p(void, ptr);
}
talloc_abort_type_mismatch(location, pname, name);
return NULL;
}
/*
this is for compatibility with older versions of talloc
*/
_PUBLIC_ void *talloc_init(const char *fmt, ...)
{
va_list ap;
void *ptr;
const char *name;
ptr = __talloc(NULL, 0);
if (unlikely(ptr == NULL)) return NULL;
va_start(ap, fmt);
name = talloc_set_name_v(ptr, fmt, ap);
va_end(ap);
if (unlikely(name == NULL)) {
_talloc_free_internal(ptr, __location__);
return NULL;
}
return ptr;
}
static inline void _talloc_free_children_internal(struct talloc_chunk *tc,
void *ptr,
const char *location)
{
while (tc->child) {
/* we need to work out who will own an abandoned child
if it cannot be freed. In priority order, the first
choice is owner of any remaining reference to this
pointer, the second choice is our parent, and the
final choice is the null context. */
void *child = TC_PTR_FROM_CHUNK(tc->child);
const void *new_parent = null_context;
if (unlikely(tc->child->refs)) {
struct talloc_chunk *p = talloc_parent_chunk(tc->child->refs);
if (p) new_parent = TC_PTR_FROM_CHUNK(p);
}
if (unlikely(_talloc_free_internal(child, location) == -1)) {
if (new_parent == null_context) {
struct talloc_chunk *p = talloc_parent_chunk(ptr);
if (p) new_parent = TC_PTR_FROM_CHUNK(p);
}
_talloc_steal_internal(new_parent, child);
}
}
}
/*
this is a replacement for the Samba3 talloc_destroy_pool functionality. It
should probably not be used in new code. It's in here to keep the talloc
code consistent across Samba 3 and 4.
*/
_PUBLIC_ void talloc_free_children(void *ptr)
{
struct talloc_chunk *tc_name = NULL;
struct talloc_chunk *tc;
if (unlikely(ptr == NULL)) {
return;
}
tc = talloc_chunk_from_ptr(ptr);
/* we do not want to free the context name if it is a child .. */
if (likely(tc->child)) {
for (tc_name = tc->child; tc_name; tc_name = tc_name->next) {
if (tc->name == TC_PTR_FROM_CHUNK(tc_name)) break;
}
if (tc_name) {
_TLIST_REMOVE(tc->child, tc_name);
if (tc->child) {
tc->child->parent = tc;
}
}
}
_talloc_free_children_internal(tc, ptr, __location__);
/* .. so we put it back after all other children have been freed */
if (tc_name) {
if (tc->child) {
tc->child->parent = NULL;
}
tc_name->parent = tc;
_TLIST_ADD(tc->child, tc_name);
}
}
/*
Allocate a bit of memory as a child of an existing pointer
*/
_PUBLIC_ void *_talloc(const void *context, size_t size)
{
return __talloc(context, size);
}
/*
externally callable talloc_set_name_const()
*/
_PUBLIC_ void talloc_set_name_const(const void *ptr, const char *name)
{
_talloc_set_name_const(ptr, name);
}
/*
create a named talloc pointer. Any talloc pointer can be named, and
talloc_named() operates just like talloc() except that it allows you
to name the pointer.
*/
_PUBLIC_ void *talloc_named_const(const void *context, size_t size, const char *name)
{
return _talloc_named_const(context, size, name);
}
/*
free a talloc pointer. This also frees all child pointers of this
pointer recursively
return 0 if the memory is actually freed, otherwise -1. The memory
will not be freed if the ref_count is > 1 or the destructor (if
any) returns non-zero
*/
_PUBLIC_ int _talloc_free(void *ptr, const char *location)
{
struct talloc_chunk *tc;
if (unlikely(ptr == NULL)) {
return -1;
}
tc = talloc_chunk_from_ptr(ptr);
if (unlikely(tc->refs != NULL)) {
struct talloc_reference_handle *h;
if (talloc_parent(ptr) == null_context && tc->refs->next == NULL) {
/* in this case we do know which parent should
get this pointer, as there is really only
one parent */
return talloc_unlink(null_context, ptr);
}
talloc_log("ERROR: talloc_free with references at %s\n",
location);
for (h=tc->refs; h; h=h->next) {
talloc_log("\treference at %s\n",
h->location);
}
return -1;
}
return _talloc_free_internal(ptr, location);
}
/*
A talloc version of realloc. The context argument is only used if
ptr is NULL
*/
_PUBLIC_ void *_talloc_realloc(const void *context, void *ptr, size_t size, const char *name)
{
struct talloc_chunk *tc;
void *new_ptr;
bool malloced = false;
union talloc_pool_chunk *pool_tc = NULL;
/* size zero is equivalent to free() */
if (unlikely(size == 0)) {
talloc_unlink(context, ptr);
return NULL;
}
if (unlikely(size >= MAX_TALLOC_SIZE)) {
return NULL;
}
/* realloc(NULL) is equivalent to malloc() */
if (ptr == NULL) {
return _talloc_named_const(context, size, name);
}
tc = talloc_chunk_from_ptr(ptr);
/* don't allow realloc on referenced pointers */
if (unlikely(tc->refs)) {
return NULL;
}
/* don't let anybody try to realloc a talloc_pool */
if (unlikely(tc->flags & TALLOC_FLAG_POOL)) {
return NULL;
}
if (tc->limit && (size - tc->size > 0)) {
if (!talloc_memlimit_check(tc->limit, (size - tc->size))) {
errno = ENOMEM;
return NULL;
}
}
/* handle realloc inside a talloc_pool */
if (unlikely(tc->flags & TALLOC_FLAG_POOLMEM)) {
pool_tc = (union talloc_pool_chunk *)tc->pool;
}
#if (ALWAYS_REALLOC == 0)
/* don't shrink if we have less than 1k to gain */
if (size < tc->size && tc->limit == NULL) {
if (pool_tc) {
void *next_tc = tc_next_chunk(tc);
TC_INVALIDATE_SHRINK_CHUNK(tc, size);
tc->size = size;
if (next_tc == pool_tc->hdr.c.pool) {
/* note: tc->size has changed, so this works */
pool_tc->hdr.c.pool = tc_next_chunk(tc);
}
return ptr;
} else if ((tc->size - size) < 1024) {
/*
* if we call TC_INVALIDATE_SHRINK_CHUNK() here
* we would need to call TC_UNDEFINE_GROW_CHUNK()
* after each realloc call, which slows down
* testing a lot :-(.
*
* That is why we only mark memory as undefined here.
*/
TC_UNDEFINE_SHRINK_CHUNK(tc, size);
/* do not shrink if we have less than 1k to gain */
tc->size = size;
return ptr;
}
} else if (tc->size == size) {
/*
* do not change the pointer if it is exactly
* the same size.
*/
return ptr;
}
#endif
/* by resetting magic we catch users of the old memory */
tc->flags |= TALLOC_FLAG_FREE;
#if ALWAYS_REALLOC
if (pool_tc) {
new_ptr = talloc_alloc_pool(tc, size + TC_HDR_SIZE);
pool_tc->hdr.object_count--;
if (new_ptr == NULL) {
new_ptr = malloc(TC_HDR_SIZE+size);
malloced = true;
}
if (new_ptr) {
memcpy(new_ptr, tc, MIN(tc->size,size) + TC_HDR_SIZE);
TC_INVALIDATE_FULL_CHUNK(tc);
}
} else {
new_ptr = malloc(size + TC_HDR_SIZE);
if (new_ptr) {
memcpy(new_ptr, tc, MIN(tc->size, size) + TC_HDR_SIZE);
free(tc);
}
}
#else
if (pool_tc) {
void *next_tc = tc_next_chunk(tc);
size_t old_chunk_size = TC_ALIGN16(TC_HDR_SIZE + tc->size);
size_t new_chunk_size = TC_ALIGN16(TC_HDR_SIZE + size);
size_t space_needed;
size_t space_left;
unsigned int chunk_count = pool_tc->hdr.object_count;
if (!(pool_tc->hdr.c.flags & TALLOC_FLAG_FREE)) {
chunk_count -= 1;
}
if (chunk_count == 1) {
/*
* optimize for the case where 'tc' is the only
* chunk in the pool.
*/
char *start = tc_pool_first_chunk(pool_tc);
space_needed = new_chunk_size;
space_left = (char *)tc_pool_end(pool_tc) - start;
if (space_left >= space_needed) {
size_t old_used = TC_HDR_SIZE + tc->size;
size_t new_used = TC_HDR_SIZE + size;
new_ptr = start;
memmove(new_ptr, tc, old_used);
tc = (struct talloc_chunk *)new_ptr;
TC_UNDEFINE_GROW_CHUNK(tc, size);
/*
* first we do not align the pool pointer
* because we want to invalidate the padding
* too.
*/
pool_tc->hdr.c.pool = new_used + (char *)new_ptr;
tc_invalidate_pool(pool_tc);
/* now the aligned pointer */
pool_tc->hdr.c.pool = new_chunk_size + (char *)new_ptr;
goto got_new_ptr;
}
next_tc = NULL;
}
if (new_chunk_size == old_chunk_size) {
TC_UNDEFINE_GROW_CHUNK(tc, size);
tc->flags &= ~TALLOC_FLAG_FREE;
if (!talloc_memlimit_update(tc->limit,
tc->size, size)) {
talloc_abort("cur_size memlimit counter not"
" correct!");
errno = EINVAL;
return NULL;
}
tc->size = size;
return ptr;
}
if (next_tc == pool_tc->hdr.c.pool) {
/*
* optimize for the case where 'tc' is the last
* chunk in the pool.
*/
space_needed = new_chunk_size - old_chunk_size;
space_left = tc_pool_space_left(pool_tc);
if (space_left >= space_needed) {
TC_UNDEFINE_GROW_CHUNK(tc, size);
tc->flags &= ~TALLOC_FLAG_FREE;
if (!talloc_memlimit_update(tc->limit,
tc->size, size)) {
talloc_abort("cur_size memlimit "
"counter not correct!");
errno = EINVAL;
return NULL;
}
tc->size = size;
pool_tc->hdr.c.pool = tc_next_chunk(tc);
return ptr;
}
}
new_ptr = talloc_alloc_pool(tc, size + TC_HDR_SIZE);
if (new_ptr == NULL) {
new_ptr = malloc(TC_HDR_SIZE+size);
malloced = true;
}
if (new_ptr) {
memcpy(new_ptr, tc, MIN(tc->size,size) + TC_HDR_SIZE);
_talloc_free_poolmem(tc, __location__ "_talloc_realloc");
}
}
else {
new_ptr = realloc(tc, size + TC_HDR_SIZE);
}
got_new_ptr:
#endif
if (unlikely(!new_ptr)) {
tc->flags &= ~TALLOC_FLAG_FREE;
return NULL;
}
tc = (struct talloc_chunk *)new_ptr;
tc->flags &= ~TALLOC_FLAG_FREE;
if (malloced) {
tc->flags &= ~TALLOC_FLAG_POOLMEM;
}
if (tc->parent) {
tc->parent->child = tc;
}
if (tc->child) {
tc->child->parent = tc;
}
if (tc->prev) {
tc->prev->next = tc;
}
if (tc->next) {
tc->next->prev = tc;
}
if (!talloc_memlimit_update(tc->limit, tc->size, size)) {
talloc_abort("cur_size memlimit counter not correct!");
errno = EINVAL;
return NULL;
}
tc->size = size;
_talloc_set_name_const(TC_PTR_FROM_CHUNK(tc), name);
return TC_PTR_FROM_CHUNK(tc);
}
/*
a wrapper around talloc_steal() for situations where you are moving a pointer
between two structures, and want the old pointer to be set to NULL
*/
_PUBLIC_ void *_talloc_move(const void *new_ctx, const void *_pptr)
{
const void **pptr = discard_const_p(const void *,_pptr);
void *ret = talloc_steal(new_ctx, discard_const_p(void, *pptr));
(*pptr) = NULL;
return ret;
}
enum talloc_mem_count_type {
TOTAL_MEM_SIZE,
TOTAL_MEM_BLOCKS,
TOTAL_MEM_LIMIT,
};
static size_t _talloc_total_mem_internal(const void *ptr,
enum talloc_mem_count_type type,
struct talloc_memlimit *old_limit,
struct talloc_memlimit *new_limit)
{
size_t total = 0;
struct talloc_chunk *c, *tc;
if (ptr == NULL) {
ptr = null_context;
}
if (ptr == NULL) {
return 0;
}
tc = talloc_chunk_from_ptr(ptr);
if (old_limit || new_limit) {
if (tc->limit && tc->limit->upper == old_limit) {
tc->limit->upper = new_limit;
}
}
/* optimize in the memlimits case */
if (type == TOTAL_MEM_LIMIT &&
tc->limit != NULL &&
tc->limit != old_limit &&
tc->limit->parent == tc) {
return tc->limit->cur_size;
}
if (tc->flags & TALLOC_FLAG_LOOP) {
return 0;
}
tc->flags |= TALLOC_FLAG_LOOP;
if (old_limit || new_limit) {
if (old_limit == tc->limit) {
tc->limit = new_limit;
}
}
switch (type) {
case TOTAL_MEM_SIZE:
if (likely(tc->name != TALLOC_MAGIC_REFERENCE)) {
total = tc->size;
}
break;
case TOTAL_MEM_BLOCKS:
total++;
break;
case TOTAL_MEM_LIMIT:
if (likely(tc->name != TALLOC_MAGIC_REFERENCE)) {
total = tc->size + TC_HDR_SIZE;
}
break;
}
for (c = tc->child; c; c = c->next) {
total += _talloc_total_mem_internal(TC_PTR_FROM_CHUNK(c), type,
old_limit, new_limit);
}
tc->flags &= ~TALLOC_FLAG_LOOP;
return total;
}
/*
return the total size of a talloc pool (subtree)
*/
_PUBLIC_ size_t talloc_total_size(const void *ptr)
{
return _talloc_total_mem_internal(ptr, TOTAL_MEM_SIZE, NULL, NULL);
}
/*
return the total number of blocks in a talloc pool (subtree)
*/
_PUBLIC_ size_t talloc_total_blocks(const void *ptr)
{
return _talloc_total_mem_internal(ptr, TOTAL_MEM_BLOCKS, NULL, NULL);
}
/*
return the number of external references to a pointer
*/
_PUBLIC_ size_t talloc_reference_count(const void *ptr)
{
struct talloc_chunk *tc = talloc_chunk_from_ptr(ptr);
struct talloc_reference_handle *h;
size_t ret = 0;
for (h=tc->refs;h;h=h->next) {
ret++;
}
return ret;
}
/*
report on memory usage by all children of a pointer, giving a full tree view
*/
_PUBLIC_ void talloc_report_depth_cb(const void *ptr, int depth, int max_depth,
void (*callback)(const void *ptr,
int depth, int max_depth,
int is_ref,
void *private_data),
void *private_data)
{
struct talloc_chunk *c, *tc;
if (ptr == NULL) {
ptr = null_context;
}
if (ptr == NULL) return;
tc = talloc_chunk_from_ptr(ptr);
if (tc->flags & TALLOC_FLAG_LOOP) {
return;
}
callback(ptr, depth, max_depth, 0, private_data);
if (max_depth >= 0 && depth >= max_depth) {
return;
}
tc->flags |= TALLOC_FLAG_LOOP;
for (c=tc->child;c;c=c->next) {
if (c->name == TALLOC_MAGIC_REFERENCE) {
struct talloc_reference_handle *h = (struct talloc_reference_handle *)TC_PTR_FROM_CHUNK(c);
callback(h->ptr, depth + 1, max_depth, 1, private_data);
} else {
talloc_report_depth_cb(TC_PTR_FROM_CHUNK(c), depth + 1, max_depth, callback, private_data);
}
}
tc->flags &= ~TALLOC_FLAG_LOOP;
}
static void talloc_report_depth_FILE_helper(const void *ptr, int depth, int max_depth, int is_ref, void *_f)
{
const char *name = talloc_get_name(ptr);
struct talloc_chunk *tc;
FILE *f = (FILE *)_f;
if (is_ref) {
fprintf(f, "%*sreference to: %s\n", depth*4, "", name);
return;
}
tc = talloc_chunk_from_ptr(ptr);
if (tc->limit && tc->limit->parent == tc) {
fprintf(f, "%*s%-30s is a memlimit context"
" (max_size = %lu bytes, cur_size = %lu bytes)\n",
depth*4, "",
name,
(unsigned long)tc->limit->max_size,
(unsigned long)tc->limit->cur_size);
}
if (depth == 0) {
fprintf(f,"%stalloc report on '%s' (total %6lu bytes in %3lu blocks)\n",
(max_depth < 0 ? "full " :""), name,
(unsigned long)talloc_total_size(ptr),
(unsigned long)talloc_total_blocks(ptr));
return;
}
fprintf(f, "%*s%-30s contains %6lu bytes in %3lu blocks (ref %d) %p\n",
depth*4, "",
name,
(unsigned long)talloc_total_size(ptr),
(unsigned long)talloc_total_blocks(ptr),
(int)talloc_reference_count(ptr), ptr);
#if 0
fprintf(f, "content: ");
if (talloc_total_size(ptr)) {
int tot = talloc_total_size(ptr);
int i;
for (i = 0; i < tot; i++) {
if ((((char *)ptr)[i] > 31) && (((char *)ptr)[i] < 126)) {
fprintf(f, "%c", ((char *)ptr)[i]);
} else {
fprintf(f, "~%02x", ((char *)ptr)[i]);
}
}
}
fprintf(f, "\n");
#endif
}
/*
report on memory usage by all children of a pointer, giving a full tree view
*/
_PUBLIC_ void talloc_report_depth_file(const void *ptr, int depth, int max_depth, FILE *f)
{
if (f) {
talloc_report_depth_cb(ptr, depth, max_depth, talloc_report_depth_FILE_helper, f);
fflush(f);
}
}
/*
report on memory usage by all children of a pointer, giving a full tree view
*/
_PUBLIC_ void talloc_report_full(const void *ptr, FILE *f)
{
talloc_report_depth_file(ptr, 0, -1, f);
}
/*
report on memory usage by all children of a pointer
*/
_PUBLIC_ void talloc_report(const void *ptr, FILE *f)
{
talloc_report_depth_file(ptr, 0, 1, f);
}
/*
report on any memory hanging off the null context
*/
static void talloc_report_null(void)
{
if (talloc_total_size(null_context) != 0) {
talloc_report(null_context, stderr);
}
}
/*
report on any memory hanging off the null context
*/
static void talloc_report_null_full(void)
{
if (talloc_total_size(null_context) != 0) {
talloc_report_full(null_context, stderr);
}
}
/*
enable tracking of the NULL context
*/
_PUBLIC_ void talloc_enable_null_tracking(void)
{
if (null_context == NULL) {
null_context = _talloc_named_const(NULL, 0, "null_context");
if (autofree_context != NULL) {
talloc_reparent(NULL, null_context, autofree_context);
}
}
}
/*
enable tracking of the NULL context, not moving the autofree context
into the NULL context. This is needed for the talloc testsuite
*/
_PUBLIC_ void talloc_enable_null_tracking_no_autofree(void)
{
if (null_context == NULL) {
null_context = _talloc_named_const(NULL, 0, "null_context");
}
}
/*
disable tracking of the NULL context
*/
_PUBLIC_ void talloc_disable_null_tracking(void)
{
if (null_context != NULL) {
/* we have to move any children onto the real NULL
context */
struct talloc_chunk *tc, *tc2;
tc = talloc_chunk_from_ptr(null_context);
for (tc2 = tc->child; tc2; tc2=tc2->next) {
if (tc2->parent == tc) tc2->parent = NULL;
if (tc2->prev == tc) tc2->prev = NULL;
}
for (tc2 = tc->next; tc2; tc2=tc2->next) {
if (tc2->parent == tc) tc2->parent = NULL;
if (tc2->prev == tc) tc2->prev = NULL;
}
tc->child = NULL;
tc->next = NULL;
}
talloc_free(null_context);
null_context = NULL;
}
/*
enable leak reporting on exit
*/
_PUBLIC_ void talloc_enable_leak_report(void)
{
talloc_enable_null_tracking();
atexit(talloc_report_null);
}
/*
enable full leak reporting on exit
*/
_PUBLIC_ void talloc_enable_leak_report_full(void)
{
talloc_enable_null_tracking();
atexit(talloc_report_null_full);
}
/*
talloc and zero memory.
*/
_PUBLIC_ void *_talloc_zero(const void *ctx, size_t size, const char *name)
{
void *p = _talloc_named_const(ctx, size, name);
if (p) {
memset(p, '\0', size);
}
return p;
}
/*
memdup with a talloc.
*/
_PUBLIC_ void *_talloc_memdup(const void *t, const void *p, size_t size, const char *name)
{
void *newp = _talloc_named_const(t, size, name);
if (likely(newp)) {
memcpy(newp, p, size);
}
return newp;
}
static inline char *__talloc_strlendup(const void *t, const char *p, size_t len)
{
char *ret;
ret = (char *)__talloc(t, len + 1);
if (unlikely(!ret)) return NULL;
memcpy(ret, p, len);
ret[len] = 0;
_talloc_set_name_const(ret, ret);
return ret;
}
/*
strdup with a talloc
*/
_PUBLIC_ char *talloc_strdup(const void *t, const char *p)
{
if (unlikely(!p)) return NULL;
return __talloc_strlendup(t, p, strlen(p));
}
/*
strndup with a talloc
*/
_PUBLIC_ char *talloc_strndup(const void *t, const char *p, size_t n)
{
if (unlikely(!p)) return NULL;
return __talloc_strlendup(t, p, strnlen(p, n));
}
static inline char *__talloc_strlendup_append(char *s, size_t slen,
const char *a, size_t alen)
{
char *ret;
ret = talloc_realloc(NULL, s, char, slen + alen + 1);
if (unlikely(!ret)) return NULL;
/* append the string and the trailing \0 */
memcpy(&ret[slen], a, alen);
ret[slen+alen] = 0;
_talloc_set_name_const(ret, ret);
return ret;
}
/*
* Appends at the end of the string.
*/
_PUBLIC_ char *talloc_strdup_append(char *s, const char *a)
{
if (unlikely(!s)) {
return talloc_strdup(NULL, a);
}
if (unlikely(!a)) {
return s;
}
return __talloc_strlendup_append(s, strlen(s), a, strlen(a));
}
/*
* Appends at the end of the talloc'ed buffer,
* not the end of the string.
*/
_PUBLIC_ char *talloc_strdup_append_buffer(char *s, const char *a)
{
size_t slen;
if (unlikely(!s)) {
return talloc_strdup(NULL, a);
}
if (unlikely(!a)) {
return s;
}
slen = talloc_get_size(s);
if (likely(slen > 0)) {
slen--;
}
return __talloc_strlendup_append(s, slen, a, strlen(a));
}
/*
* Appends at the end of the string.
*/
_PUBLIC_ char *talloc_strndup_append(char *s, const char *a, size_t n)
{
if (unlikely(!s)) {
return talloc_strndup(NULL, a, n);
}
if (unlikely(!a)) {
return s;
}
return __talloc_strlendup_append(s, strlen(s), a, strnlen(a, n));
}
/*
* Appends at the end of the talloc'ed buffer,
* not the end of the string.
*/
_PUBLIC_ char *talloc_strndup_append_buffer(char *s, const char *a, size_t n)
{
size_t slen;
if (unlikely(!s)) {
return talloc_strndup(NULL, a, n);
}
if (unlikely(!a)) {
return s;
}
slen = talloc_get_size(s);
if (likely(slen > 0)) {
slen--;
}
return __talloc_strlendup_append(s, slen, a, strnlen(a, n));
}
#ifndef HAVE_VA_COPY
#ifdef HAVE___VA_COPY
#define va_copy(dest, src) __va_copy(dest, src)
#else
#define va_copy(dest, src) (dest) = (src)
#endif
#endif
_PUBLIC_ char *talloc_vasprintf(const void *t, const char *fmt, va_list ap)
{
int len;
char *ret;
va_list ap2;
char c;
/* this call looks strange, but it makes it work on older solaris boxes */
va_copy(ap2, ap);
len = vsnprintf(&c, 1, fmt, ap2);
va_end(ap2);
if (unlikely(len < 0)) {
return NULL;
}
ret = (char *)__talloc(t, len+1);
if (unlikely(!ret)) return NULL;
va_copy(ap2, ap);
vsnprintf(ret, len+1, fmt, ap2);
va_end(ap2);
_talloc_set_name_const(ret, ret);
return ret;
}
/*
Perform string formatting, and return a pointer to newly allocated
memory holding the result, inside a memory pool.
*/
_PUBLIC_ char *talloc_asprintf(const void *t, const char *fmt, ...)
{
va_list ap;
char *ret;
va_start(ap, fmt);
ret = talloc_vasprintf(t, fmt, ap);
va_end(ap);
return ret;
}
static inline char *__talloc_vaslenprintf_append(char *s, size_t slen,
const char *fmt, va_list ap)
PRINTF_ATTRIBUTE(3,0);
static inline char *__talloc_vaslenprintf_append(char *s, size_t slen,
const char *fmt, va_list ap)
{
ssize_t alen;
va_list ap2;
char c;
va_copy(ap2, ap);
alen = vsnprintf(&c, 1, fmt, ap2);
va_end(ap2);
if (alen <= 0) {
/* Either the vsnprintf failed or the format resulted in
* no characters being formatted. In the former case, we
* ought to return NULL, in the latter we ought to return
* the original string. Most current callers of this
* function expect it to never return NULL.
*/
return s;
}
s = talloc_realloc(NULL, s, char, slen + alen + 1);
if (!s) return NULL;
va_copy(ap2, ap);
vsnprintf(s + slen, alen + 1, fmt, ap2);
va_end(ap2);
_talloc_set_name_const(s, s);
return s;
}
/**
* Realloc @p s to append the formatted result of @p fmt and @p ap,
* and return @p s, which may have moved. Good for gradually
* accumulating output into a string buffer. Appends at the end
* of the string.
**/
_PUBLIC_ char *talloc_vasprintf_append(char *s, const char *fmt, va_list ap)
{
if (unlikely(!s)) {
return talloc_vasprintf(NULL, fmt, ap);
}
return __talloc_vaslenprintf_append(s, strlen(s), fmt, ap);
}
/**
* Realloc @p s to append the formatted result of @p fmt and @p ap,
* and return @p s, which may have moved. Always appends at the
* end of the talloc'ed buffer, not the end of the string.
**/
_PUBLIC_ char *talloc_vasprintf_append_buffer(char *s, const char *fmt, va_list ap)
{
size_t slen;
if (unlikely(!s)) {
return talloc_vasprintf(NULL, fmt, ap);
}
slen = talloc_get_size(s);
if (likely(slen > 0)) {
slen--;
}
return __talloc_vaslenprintf_append(s, slen, fmt, ap);
}
/*
Realloc @p s to append the formatted result of @p fmt and return @p
s, which may have moved. Good for gradually accumulating output
into a string buffer.
*/
_PUBLIC_ char *talloc_asprintf_append(char *s, const char *fmt, ...)
{
va_list ap;
va_start(ap, fmt);
s = talloc_vasprintf_append(s, fmt, ap);
va_end(ap);
return s;
}
/*
Realloc @p s to append the formatted result of @p fmt and return @p
s, which may have moved. Good for gradually accumulating output
into a buffer.
*/
_PUBLIC_ char *talloc_asprintf_append_buffer(char *s, const char *fmt, ...)
{
va_list ap;
va_start(ap, fmt);
s = talloc_vasprintf_append_buffer(s, fmt, ap);
va_end(ap);
return s;
}
/*
alloc an array, checking for integer overflow in the array size
*/
_PUBLIC_ void *_talloc_array(const void *ctx, size_t el_size, unsigned count, const char *name)
{
if (count >= MAX_TALLOC_SIZE/el_size) {
return NULL;
}
return _talloc_named_const(ctx, el_size * count, name);
}
/*
alloc an zero array, checking for integer overflow in the array size
*/
_PUBLIC_ void *_talloc_zero_array(const void *ctx, size_t el_size, unsigned count, const char *name)
{
if (count >= MAX_TALLOC_SIZE/el_size) {
return NULL;
}
return _talloc_zero(ctx, el_size * count, name);
}
/*
realloc an array, checking for integer overflow in the array size
*/
_PUBLIC_ void *_talloc_realloc_array(const void *ctx, void *ptr, size_t el_size, unsigned count, const char *name)
{
if (count >= MAX_TALLOC_SIZE/el_size) {
return NULL;
}
return _talloc_realloc(ctx, ptr, el_size * count, name);
}
/*
a function version of talloc_realloc(), so it can be passed as a function pointer
to libraries that want a realloc function (a realloc function encapsulates
all the basic capabilities of an allocation library, which is why this is useful)
*/
_PUBLIC_ void *talloc_realloc_fn(const void *context, void *ptr, size_t size)
{
return _talloc_realloc(context, ptr, size, NULL);
}
static int talloc_autofree_destructor(void *ptr)
{
autofree_context = NULL;
return 0;
}
static void talloc_autofree(void)
{
talloc_free(autofree_context);
}
/*
return a context which will be auto-freed on exit
this is useful for reducing the noise in leak reports
*/
_PUBLIC_ void *talloc_autofree_context(void)
{
if (autofree_context == NULL) {
autofree_context = _talloc_named_const(NULL, 0, "autofree_context");
talloc_set_destructor(autofree_context, talloc_autofree_destructor);
atexit(talloc_autofree);
}
return autofree_context;
}
_PUBLIC_ size_t talloc_get_size(const void *context)
{
struct talloc_chunk *tc;
if (context == NULL) {
context = null_context;
}
if (context == NULL) {
return 0;
}
tc = talloc_chunk_from_ptr(context);
return tc->size;
}
/*
find a parent of this context that has the given name, if any
*/
_PUBLIC_ void *talloc_find_parent_byname(const void *context, const char *name)
{
struct talloc_chunk *tc;
if (context == NULL) {
return NULL;
}
tc = talloc_chunk_from_ptr(context);
while (tc) {
if (tc->name && strcmp(tc->name, name) == 0) {
return TC_PTR_FROM_CHUNK(tc);
}
while (tc && tc->prev) tc = tc->prev;
if (tc) {
tc = tc->parent;
}
}
return NULL;
}
/*
show the parentage of a context
*/
_PUBLIC_ void talloc_show_parents(const void *context, FILE *file)
{
struct talloc_chunk *tc;
if (context == NULL) {
fprintf(file, "talloc no parents for NULL\n");
return;
}
tc = talloc_chunk_from_ptr(context);
fprintf(file, "talloc parents of '%s'\n", talloc_get_name(context));
while (tc) {
fprintf(file, "\t'%s'\n", talloc_get_name(TC_PTR_FROM_CHUNK(tc)));
while (tc && tc->prev) tc = tc->prev;
if (tc) {
tc = tc->parent;
}
}
fflush(file);
}
/*
return 1 if ptr is a parent of context
*/
static int _talloc_is_parent(const void *context, const void *ptr, int depth)
{
struct talloc_chunk *tc;
if (context == NULL) {
return 0;
}
tc = talloc_chunk_from_ptr(context);
while (tc && depth > 0) {
if (TC_PTR_FROM_CHUNK(tc) == ptr) return 1;
while (tc && tc->prev) tc = tc->prev;
if (tc) {
tc = tc->parent;
depth--;
}
}
return 0;
}
/*
return 1 if ptr is a parent of context
*/
_PUBLIC_ int talloc_is_parent(const void *context, const void *ptr)
{
return _talloc_is_parent(context, ptr, TALLOC_MAX_DEPTH);
}
/*
return the total size of memory used by this context and all children
*/
static size_t _talloc_total_limit_size(const void *ptr,
struct talloc_memlimit *old_limit,
struct talloc_memlimit *new_limit)
{
return _talloc_total_mem_internal(ptr, TOTAL_MEM_LIMIT,
old_limit, new_limit);
}
static bool talloc_memlimit_check(struct talloc_memlimit *limit, size_t size)
{
struct talloc_memlimit *l;
for (l = limit; l != NULL; l = l->upper) {
if (l->max_size != 0 &&
((l->max_size <= l->cur_size) ||
(l->max_size - l->cur_size < TC_HDR_SIZE+size))) {
return false;
}
}
return true;
}
static bool talloc_memlimit_update(struct talloc_memlimit *limit,
size_t old_size, size_t new_size)
{
struct talloc_memlimit *l;
ssize_t d;
if (old_size == 0) {
d = new_size + TC_HDR_SIZE;
} else {
d = new_size - old_size;
}
for (l = limit; l != NULL; l = l->upper) {
ssize_t new_cur_size = l->cur_size + d;
if (new_cur_size < 0) {
return false;
}
l->cur_size = new_cur_size;
}
return true;
}
_PUBLIC_ int talloc_set_memlimit(const void *ctx, size_t max_size)
{
struct talloc_chunk *tc = talloc_chunk_from_ptr(ctx);
struct talloc_memlimit *orig_limit;
struct talloc_memlimit *limit = NULL;
if (tc->limit && tc->limit->parent == tc) {
tc->limit->max_size = max_size;
return 0;
}
orig_limit = tc->limit;
limit = malloc(sizeof(struct talloc_memlimit));
if (limit == NULL) {
return 1;
}
limit->parent = tc;
limit->max_size = max_size;
limit->cur_size = _talloc_total_limit_size(ctx, tc->limit, limit);
if (orig_limit) {
limit->upper = orig_limit;
} else {
limit->upper = NULL;
}
return 0;
}
f='#n1660'>1660
1661
1662
1663
1664
1665
1666
1667
1668
1669
1670
1671
1672
1673
1674
1675
1676
1677
1678
1679
1680
1681
1682
1683
1684
1685
1686
1687
1688
1689
1690
1691
1692
1693
1694
1695
1696
1697
1698
1699
1700
1701
1702
1703
1704
1705
1706
1707
1708
1709
1710
1711
1712
1713
1714
1715
1716
1717
1718
1719
1720
1721
1722
1723
1724
1725
1726
1727
1728
1729
1730
1731
1732
1733
1734
1735
1736
1737
1738
1739
1740
1741
1742
1743
1744
1745
1746
1747
1748
1749
1750
1751
1752
1753
1754
1755
1756
1757
1758
1759
1760
1761
1762
1763
1764
1765
1766
1767
1768
1769
1770
1771
1772
1773
1774
1775
1776
1777
1778
1779
1780
1781
1782
1783
1784
1785
1786
1787
1788
1789
1790
1791
1792
1793
1794
1795
1796
1797
1798
1799
1800
1801
1802
1803
1804
1805
1806
1807
1808
1809
1810
1811
1812
1813
1814
1815
1816
1817
1818
1819
1820
1821
1822
1823
1824
1825
1826
1827
1828
1829
1830
1831
1832
1833
1834
1835
1836
1837
1838
1839
1840
1841
1842
1843
1844
1845
1846
1847
1848
1849
1850
1851
1852
1853
1854
1855
1856
1857
1858
1859
1860
1861
1862
1863
1864
1865
1866
1867
1868
1869
1870
1871
1872
1873
1874
1875
1876
1877
1878
1879
1880
1881
1882
1883
1884
1885
1886
1887
1888
1889
1890
1891
1892
1893
1894
1895
1896
1897
1898
1899
1900
1901
1902
1903
1904
1905
1906
1907
1908
1909
1910
1911
1912
1913
1914
1915
1916
1917
1918
1919
1920
1921
1922
1923
1924
1925
1926
1927
1928
1929
1930
1931
1932
1933
1934
1935
1936
1937
1938
1939
1940
1941
1942
1943
1944
1945
1946
1947
1948
1949
1950
1951
1952
1953
1954
1955
1956
1957
1958
1959
1960
1961
1962
1963
1964
1965
1966
1967
1968
1969
1970
1971
1972
1973
1974
1975
1976
1977
1978
1979
1980
1981
1982
1983
1984
1985
1986
1987
1988
1989
1990
1991
1992
1993
1994
1995
1996
1997
1998
1999
2000
2001
2002
2003
2004
2005
2006
2007
2008
2009
2010
2011
2012
2013
2014
2015
2016
2017
2018
2019
2020
2021
2022
2023
2024
2025
2026
2027
2028
2029
2030
2031
2032
2033
2034
2035
2036
2037
2038
2039
2040
2041
2042
2043
2044
2045
2046
2047
2048
2049
2050
2051
2052
2053
2054
2055
2056
2057
2058
2059
2060
2061
2062
2063
2064
2065
2066
2067
2068
2069
2070
2071
2072
2073
2074
2075
2076
2077
2078
2079
2080
2081
2082
2083
2084
/*
* appl/bsd/krshd.c
*/
/*
* Copyright (c) 1983 The Regents of the University of California.
* All rights reserved.
*
* Redistribution and use in source and binary forms are permitted
* provided that the above copyright notice and this paragraph are
* duplicated in all such forms and that any documentation,
* advertising materials, and other materials related to such
* distribution and use acknowledge that the software was developed
* by the University of California, Berkeley. The name of the
* University may not be used to endorse or promote products derived
* from this software without specific prior written permission.
* THIS SOFTWARE IS PROVIDED ``AS IS'' AND WITHOUT ANY EXPRESS OR
* IMPLIED WARRANTIES, INCLUDING, WITHOUT LIMITATION, THE IMPLIED
* WARRANTIES OF MERCHANTIBILITY AND FITNESS FOR A PARTICULAR PURPOSE.
*/
#ifndef lint
char copyright[] =
"@(#) Copyright (c) 1983 The Regents of the University of California.\n\
All rights reserved.\n";
#endif /* not lint */
/* based on @(#)rshd.c 5.12 (Berkeley) 9/12/88 */
/*
* remote shell server:
* remuser\0
* locuser\0
* command\0
* data
*/
/*
* This is the rshell daemon. The very basic protocol for checking
* authentication and authorization is:
* 1) Check authentication.
* 2) Check authorization via the access-control files:
* ~/.k5login (using krb5_kuserok) and/or
* Execute command if configured authoriztion checks pass, else deny
* permission.
*
* The configuration is done either by command-line arguments passed by inetd,
* or by the name of the daemon. If command-line arguments are present, they
* take priority. The options are:
* -k means trust krb4 or krb5
* -5 means trust krb5
* -4 means trust krb4 (using .klogin)
*
*/
/* DEFINES:
* KERBEROS - Define this if application is to be kerberised.
* KRB5_KRB4_COMPAT - Define this if v4 rlogin clients are also to be served.
* ALWAYS_V5_KUSEROK - Define this if you want .k5login to be
* checked even for v4 clients (instead of .klogin).
* LOG_ALL_LOGINS - Define this if you want to log all logins.
* LOG_OTHER_USERS - Define this if you want to log all principals that do
* not map onto the local user.
* LOG_REMOTE_REALM - Define this if you want to log all principals from
* remote realms.
* LOG_CMD - Define this if you want to log not only the user but also the
* command executed. This only decides the type of information
* logged. Whether or not to log is still decided by the above
* three DEFINES.
* Note: Root account access is always logged.
*/
#define SERVE_NON_KRB
#define LOG_REMOTE_REALM
#define LOG_CMD
#ifdef HAVE_UNISTD_H
#include <unistd.h>
#endif
#ifdef HAVE_STDLIB_H
#include <stdlib.h>
#endif
#ifdef __SCO__
#include <sys/unistd.h>
#endif
#include <sys/types.h>
#include <sys/ioctl.h>
#include <sys/param.h>
#if !defined(KERBEROS) || !defined(KRB5_KRB4_COMPAT)
/* Ultrix doesn't protect it vs multiple inclusion, and krb.h includes it */
#include <sys/socket.h>
#endif
#include <sys/file.h>
#include <sys/stat.h>
#include <sys/time.h>
#include <fcntl.h>
#ifdef HAVE_SYS_SELECT_H
#include <sys/select.h>
#endif
#include <netinet/in.h>
#include <arpa/inet.h>
#include <stdio.h>
#include <grp.h>
#include <errno.h>
#include <pwd.h>
#include <ctype.h>
#include <string.h>
#include <libpty.h>
#include <sys/wait.h>
#ifdef HAVE_SYS_LABEL_H
/* only SunOS 4? */
#include <sys/label.h>
#include <sys/audit.h>
#include <pwdadj.h>
#endif
#ifdef HAVE_STDARG_H
#include <stdarg.h>
#else
#include <varargs.h>
#endif
#include <signal.h>
#if !defined(KERBEROS) || !defined(KRB5_KRB4_COMPAT)
/* Ultrix doesn't protect it vs multiple inclusion, and krb.h includes it */
#include <netdb.h>
#endif
#ifdef CRAY
#ifndef NO_UDB
#include <udb.h>
#endif /* !NO_UDB */
#include <sys/category.h>
#include <netinet/ip.h>
#include <sys/tfm.h>
#include <sys/nal.h>
#include <sys/secparm.h>
#include <sys/usrv.h>
#include <sys/utsname.h>
#include <sys/sysv.h>
#include <sys/slrec.h>
#include <sys/unistd.h>
#include <path.h>
#endif /* CRAY */
#include <syslog.h>
#ifdef POSIX_TERMIOS
#include <termios.h>
#endif
#ifdef HAVE_SYS_FILIO_H
/* get FIONBIO from sys/filio.h, so what if it is a compatibility feature */
#include <sys/filio.h>
#endif
#ifdef KERBEROS
#include <krb5.h>
#include <com_err.h>
#include "loginpaths.h"
#ifdef KRB5_KRB4_COMPAT
#include <kerberosIV/krb.h>
Key_schedule v4_schedule;
#endif
#include <k5-util.h>
#ifdef HAVE_PATHS_H
#include <paths.h>
#endif
#if defined(_PATH_NOLOGIN)
#define NOLOGIN _PATH_NOLOGIN
#else
#define NOLOGIN "/etc/nologin"
#endif
#include "defines.h"
#if HAVE_ARPA_NAMESER_H
#include <arpa/nameser.h>
#endif
#ifndef MAXDNAME
#define MAXDNAME 256 /*per the rfc*/
#endif
#define ARGSTR "ek54ciD:S:M:AP:?L:w:"
#define MAXRETRIES 4
krb5_context bsd_context;
char *srvtab = NULL;
krb5_keytab keytab = NULL;
krb5_ccache ccache = NULL;
void fatal(int, const char *);
int require_encrypt = 0;
int do_encrypt = 0;
int anyport = 0;
char *kprogdir = KPROGDIR;
int netf;
int maxhostlen = 0;
int stripdomain = 1;
int always_ip = 0;
static krb5_error_code recvauth(int netfd, struct sockaddr *peersin,
int *valid_checksum);
#else /* !KERBEROS */
#define ARGSTR "RD:?"
#endif /* KERBEROS */
#ifndef HAVE_KILLPG
#define killpg(pid, sig) kill(-(pid), (sig))
#endif
/* There are two authentication related masks:
* auth_ok and auth_sent.
* The auth_ok mask is the oring of authentication systems any one
* of which can be used.
* The auth_sent mask is the oring of one or more authentication/authorization
* systems that succeeded. If the anding
* of these two masks is true, then authorization is successful.
*/
#define AUTH_KRB4 (0x1)
#define AUTH_KRB5 (0x2)
int auth_ok = 0, auth_sent = 0;
int checksum_required = 0, checksum_ignored = 0;
char *progname;
#define MAX_PROG_NAME 10
/* Leave room for 4 environment variables to be passed */
#define MAXENV 4
#define SAVEENVPAD 0,0,0,0 /* padding for envinit slots */
char *save_env[MAXENV];
int num_env = 0;
#ifdef CRAY
int secflag;
extern
#endif /* CRAY */
void error (char *fmt, ...)
#if !defined (__cplusplus) && (__GNUC__ > 2 || (__GNUC__ == 2 && __GNUC_MINOR__ >= 7))
__attribute__ ((__format__ (__printf__, 1, 2)))
#endif
;
void usage(void), getstr(int, char *, int, char *),
doit(int, struct sockaddr *);
#ifndef HAVE_INITGROUPS
int initgroups(char* name, gid_t basegid) {
gid_t others[NGROUPS_MAX+1];
int ngrps;
others[0] = basegid;
ngrps = getgroups(NGROUPS_MAX, others+1);
return setgroups(ngrps+1, others);
}
#endif
int main(argc, argv)
int argc;
char **argv;
{
#if defined(BSD) && BSD+0 >= 43
struct linger linger;
#endif
int on = 1;
socklen_t fromlen;
struct sockaddr_storage from;
extern int opterr, optind;
extern char *optarg;
int ch;
int fd;
int debug_port = 0;
#ifdef KERBEROS
krb5_error_code status;
#endif
#ifdef CRAY
secflag = sysconf(_SC_CRAY_SECURE_SYS);
#endif
progname = strrchr (*argv, '/');
progname = progname ? progname + 1 : *argv;
#ifndef LOG_ODELAY /* 4.2 syslog */
openlog(progname, LOG_PID);
#else
#ifndef LOG_AUTH
#define LOG_AUTH 0
#endif
openlog(progname, LOG_PID | LOG_ODELAY, LOG_AUTH);
#endif /* 4.2 syslog */
#ifdef KERBEROS
status = krb5_init_context(&bsd_context);
if (status) {
syslog(LOG_ERR, "Error initializing krb5: %s",
error_message(status));
exit(1);
}
#endif
/* Analyze parameters. */
opterr = 0;
while ((ch = getopt(argc, argv, ARGSTR)) != -1)
switch (ch) {
#ifdef KERBEROS
case 'k':
#ifdef KRB5_KRB4_COMPAT
auth_ok |= (AUTH_KRB5|AUTH_KRB4);
#else
auth_ok |= AUTH_KRB5;
#endif /* KRB5_KRB4_COMPAT*/
break;
case '5':
auth_ok |= AUTH_KRB5;
break;
case 'c':
checksum_required = 1;
break;
case 'i':
checksum_ignored = 1;
break;
#ifdef KRB5_KRB4_COMPAT
case '4':
auth_ok |= AUTH_KRB4;
break;
#endif
case 'e':
require_encrypt = 1;
break;
case 'S':
if ((status = krb5_kt_resolve(bsd_context, optarg, &keytab))) {
com_err(progname, status, "while resolving srvtab file %s",
optarg);
exit(2);
}
break;
case 'M':
krb5_set_default_realm(bsd_context, optarg);
break;
case 'A':
anyport = 1;
break;
case 'P':
kprogdir = optarg;
break;
case 'L':
if (num_env < MAXENV) {
save_env[num_env] = strdup(optarg);
if(!save_env[num_env++]) {
com_err(progname, ENOMEM, "in saving environment");
exit(2);
}
} else {
fprintf(stderr, "%s: Only %d -L arguments allowed\n",
progname, MAXENV);
exit(2);
}
break;
#endif
case 'D':
debug_port = atoi(optarg);
break;
case 'w':
if (!strcmp(optarg, "ip"))
always_ip = 1;
else {
char *cp;
cp = strchr(optarg, ',');
if (cp == NULL)
maxhostlen = atoi(optarg);
else if (*(++cp)) {
if (!strcmp(cp, "striplocal"))
stripdomain = 1;
else if (!strcmp(cp, "nostriplocal"))
stripdomain = 0;
else {
usage();
exit(1);
}
*(--cp) = '\0';
maxhostlen = atoi(optarg);
}
}
break;
case '?':
default:
usage();
exit(1);
break;
}
if (optind == 0) {
usage();
exit(1);
}
argc -= optind;
argv += optind;
fromlen = sizeof (from);
if (debug_port)
fd = accept_a_connection(debug_port, (struct sockaddr *)&from,
&fromlen);
else {
if (getpeername(0, (struct sockaddr *)&from, &fromlen) < 0) {
fprintf(stderr, "%s: ", progname);
perror("getpeername");
_exit(1);
}
fd = 0;
}
/*
* AIX passes an IPv4-mapped IPv6 address back from getpeername, but if
* that address is later used in connect(), it returns an error. Convert
* IPv4-mapped IPv6 addresses to simple IPv4 addresses on AIX (but don't
* do this everywhere since it isn't always the right thing to do, just
* the least wrong on AIX).
*/
#if defined(_AIX) && defined(KRB5_USE_INET6)
if (((struct sockaddr*)&from)->sa_family == AF_INET6 && IN6_IS_ADDR_V4MAPPED(&sa2sin6(&from)->sin6_addr)) {
sa2sin(&from)->sin_len = sizeof(struct sockaddr_in);
sa2sin(&from)->sin_family = AF_INET;
sa2sin(&from)->sin_port = sa2sin6(&from)->sin6_port;
memmove(&(sa2sin(&from)->sin_addr.s_addr), &(sa2sin6(&from)->sin6_addr.u6_addr.u6_addr8[12]), 4);
}
#endif
if (setsockopt(fd, SOL_SOCKET, SO_KEEPALIVE, (char *)&on,
sizeof (on)) < 0)
syslog(LOG_WARNING, "setsockopt (SO_KEEPALIVE): %m");
#if defined(BSD) && BSD+0 >= 43
linger.l_onoff = 1;
linger.l_linger = 60; /* XXX */
if (setsockopt(fd, SOL_SOCKET, SO_LINGER, (char *)&linger,
sizeof (linger)) < 0)
syslog(LOG_WARNING , "setsockopt (SO_LINGER): %m");
#endif
if (checksum_required&&checksum_ignored) {
syslog(LOG_CRIT, "Checksums are required and ignored; these options are mutually exclusive--check the documentation.");
fatal(fd, "Configuration error: mutually exclusive options specified");
}
doit(dup(fd), (struct sockaddr *) &from);
return 0;
}
#ifdef CRAY
char username[32] = "LOGNAME=";
#include <tmpdir.h>
char tmpdir[64] = "TMPDIR=";
#else
char username[20] = "USER=";
#endif
char homedir[64] = "HOME=";
char shell[64] = "SHELL=";
char term[64] = "TERM=network";
char path_rest[] = RPATH;
char remote_addr[64+NI_MAXHOST]; /* = "KRB5REMOTEADDR=" */
char remote_port[64+NI_MAXSERV]; /* = "KRB5REMOTEPORT=" */
char local_addr[64+NI_MAXHOST]; /* = "KRB5LOCALADDR=" */
char local_port[64+NI_MAXSERV]; /* = "KRB5LOCALPORT=" */
#define ADDRPAD 0,0,0,0
#define KRBPAD 0 /* KRB5CCNAME, optional */
/* The following include extra space for TZ and MAXENV pointers... */
#define COMMONVARS homedir, shell, 0/*path*/, username, term
#ifdef CRAY
char *envinit[] =
{COMMONVARS, "TZ=GMT0", tmpdir, SAVEENVPAD, KRBPAD, ADDRPAD, 0};
#define TMPDIRENV 6
char *getenv();
#else /* CRAY */
#ifdef KERBEROS
char *envinit[] =
{COMMONVARS, 0/*tz*/, SAVEENVPAD, KRBPAD, ADDRPAD, 0};
#else /* KERBEROS */
char *envinit[] =
{COMMONVARS, 0/*tz*/, SAVEENVPAD, ADDRPAD, 0};
#endif /* KERBEROS */
#endif /* CRAY */
#define TZENV 5
#define PATHENV 2
extern char **environ;
char ttyn[12]; /* Line string for wtmp entries */
#ifdef CRAY
#define SIZEOF_INADDR SIZEOF_in_addr
int maxlogs;
#else
#define SIZEOF_INADDR sizeof(struct in_addr)
#endif
#ifndef NCARGS
/* linux doesn't seem to have it... */
#define NCARGS 1024
#endif
#define NMAX 16
int pid;
char locuser[NMAX+1];
char remuser[NMAX +1];
char cmdbuf[NCARGS+1];
char *kremuser;
krb5_principal client;
krb5_authenticator *kdata;
#ifdef KRB5_KRB4_COMPAT
AUTH_DAT *v4_kdata;
KTEXT v4_ticket;
#endif
int auth_sys = 0; /* Which version of Kerberos used to authenticate */
#define KRB5_RECVAUTH_V4 4
#define KRB5_RECVAUTH_V5 5
static void
ignore_signals()
{
#ifdef POSIX_SIGNALS
struct sigaction sa;
(void)sigemptyset(&sa.sa_mask);
sa.sa_flags = 0;
sa.sa_handler = SIG_IGN;
(void)sigaction(SIGINT, &sa, (struct sigaction *)0);
(void)sigaction(SIGQUIT, &sa, (struct sigaction *)0);
(void)sigaction(SIGTERM, &sa, (struct sigaction *)0);
(void)sigaction(SIGPIPE, &sa, (struct sigaction *)0);
(void)sigaction(SIGHUP, &sa, (struct sigaction *)0);
(void)kill(-pid, SIGTERM);
#else
signal(SIGINT, SIG_IGN);
signal(SIGQUIT, SIG_IGN);
signal(SIGTERM, SIG_IGN);
signal(SIGPIPE, SIG_IGN);
signal(SIGHUP, SIG_IGN);
killpg(pid, SIGTERM);
#endif
}
static krb5_sigtype
cleanup(signumber)
int signumber;
{
ignore_signals();
wait(0);
pty_logwtmp(ttyn,"","");
syslog(LOG_INFO ,"Daemon terminated via signal %d.", signumber);
if (ccache)
krb5_cc_destroy(bsd_context, ccache);
exit(0);
}
void doit(f, fromp)
int f;
struct sockaddr *fromp;
{
char *cp;
#ifdef KERBEROS
krb5_error_code status;
#endif
int valid_checksum;
int cnt;
char *crypt();
struct passwd *pwd;
char *path;
#ifdef CRAY
#ifndef NO_UDB
struct udb *ue;
struct udb ue_static;
extern struct udb *getudbnam();
#endif
extern struct passwd *getpwnam(), *getpwuid();
static int jid;
int error();
int paddr;
struct nal nal;
int nal_error;
struct usrv usrv;
struct sysv sysv;
char *makejtmp(), *jtmpnam = 0;
int packet_level; /* Packet classification level */
long packet_compart; /* Packet compartments */
#endif /* CRAY */
int s = -1;
char hostname[NI_MAXHOST];
char *sane_host;
char hostaddra[NI_MAXHOST];
int aierr;
short port;
int pv[2], pw[2], px[2], cc;
fd_set ready, readfrom;
char buf[RCMD_BUFSIZ], sig;
struct sockaddr_storage localaddr;
#ifdef POSIX_SIGNALS
struct sigaction sa;
#endif
#ifdef IP_TOS
/* solaris has IP_TOS, but only IPTOS_* values */
#ifdef HAVE_GETTOSBYNAME
struct tosent *tp;
if ((tp = gettosbyname("interactive", "tcp")) &&
(setsockopt(f, IPPROTO_IP, IP_TOS, &tp->t_tos, sizeof(int)) < 0))
#ifdef TOS_WARN
syslog(LOG_NOTICE, "setsockopt (IP_TOS): %m");
#else
; /* silently ignore TOS errors in 6E */
#endif
#endif
#endif /* IP_TOS */
{
socklen_t sin_len = sizeof (localaddr);
if (getsockname(f, (struct sockaddr*)&localaddr, &sin_len) < 0) {
perror("getsockname");
exit(1);
}
}
#ifdef POSIX_SIGNALS
(void)sigemptyset(&sa.sa_mask);
sa.sa_flags = 0;
sa.sa_handler = SIG_DFL;
(void)sigaction(SIGINT, &sa, (struct sigaction *)0);
(void)sigaction(SIGQUIT, &sa, (struct sigaction *)0);
(void)sigaction(SIGTERM, &sa, (struct sigaction *)0);
#else
signal(SIGINT, SIG_DFL);
signal(SIGQUIT, SIG_DFL);
signal(SIGTERM, SIG_DFL);
#endif
#ifdef DEBUG
{ int t = open("/dev/tty", 2);
if (t >= 0) {
ioctl(t, TIOCNOTTY, (char *)0);
(void) close(t);
}
}
#endif
if (fromp->sa_family != AF_INET
#if defined(KRB5_USE_INET6) && defined(KERBEROS)
&& fromp->sa_family != AF_INET6
#endif
) {
syslog(LOG_ERR , "malformed from address\n");
exit(1);
}
#ifdef KERBEROS
netf = f;
#else
{
struct sockaddr_in *frompin = sa2sin(fromp);
frompin->sin_port = ntohs((u_short)frompin->sin_port);
if (frompin->sin_port >= IPPORT_RESERVED ||
frompin->sin_port < IPPORT_RESERVED/2) {
syslog(LOG_ERR , "connection from bad port\n");
exit(1);
}
}
#endif /* KERBEROS */
#ifdef CRAY
/* If this is a secure system then get the packet classification
of f. ( Note IP_SECURITY is checked in get_packet_classification:
if it's not set then the user's (root) default
classification level and compartments are returned. )
Then set this process to that level/compart so that the stderr
connection will be labeled appropriately.
*/
if (secflag) {
if (get_packet_classification(f,getuid(),
&packet_level,&packet_compart) < 0) {
syslog(LOG_ERR, "cannot get ip packet level\n");
exit(1);
}
if(secflag == TFM_UDB_5) {
if(setucmp(packet_compart, C_PROC) != 0) {
error("Unable to setucmp.\n");
exit(1);
}
} else if(secflag == TFM_UDB_6) {
if(setulvl(packet_level,C_PROC) != 0) {
error("Unable to setulvl.\n");
exit(1);
}
if(setucmp(packet_compart, C_PROC) != 0) {
error("Unable to setucmp.\n");
exit(1);
}
}
}
#endif /* CRAY */
(void) alarm(60);
port = 0;
for (;;) {
char c;
if ((cc = read(f, &c, 1)) != 1) {
if (cc < 0)
syslog(LOG_NOTICE , "read: %m");
shutdown(f, 1+1);
exit(1);
}
if (c == 0)
break;
port = port * 10 + c - '0';
}
(void) alarm(0);
if (port != 0) {
if (anyport) {
int addrfamily = fromp->sa_family;
s = getport(0, &addrfamily);
} else {
int lport = IPPORT_RESERVED - 1;
#ifdef HAVE_RRESVPORT_AF
s = rresvport_af(&lport, fromp->sa_family);
#else
s = rresvport(&lport);
#endif
}
if (s < 0) {
syslog(LOG_ERR ,
"can't get stderr port: %m");
exit(1);
}
#ifndef KERBEROS
if (port >= IPPORT_RESERVED) {
|