/* * Copyright (C) 2011 STRATO AG * written by Arne Jansen <sensille@gmx.net> * Distributed under the GNU GPL license version 2. */ #include <linux/slab.h> #include <linux/export.h> #include "ulist.h" /* * ulist is a generic data structure to hold a collection of unique u64 * values. The only operations it supports is adding to the list and * enumerating it. * It is possible to store an auxiliary value along with the key. * * The implementation is preliminary and can probably be sped up * significantly. A first step would be to store the values in an rbtree * as soon as ULIST_SIZE is exceeded. * * A sample usage for ulists is the enumeration of directed graphs without * visiting a node twice. The pseudo-code could look like this: * * ulist = ulist_alloc(); * ulist_add(ulist, root); * ULIST_ITER_INIT(&uiter); * * while ((elem = ulist_next(ulist, &uiter)) { * for (all child nodes n in elem) * ulist_add(ulist, n); * do something useful with the node; * } * ulist_free(ulist); * * This assumes the graph nodes are adressable by u64. This stems from the * usage for tree enumeration in btrfs, where the logical addresses are * 64 bit. * * It is also useful for tree enumeration which could be done elegantly * recursively, but is not possible due to kernel stack limitations. The * loop would be similar to the above. */ /** * ulist_init - freshly initialize a ulist * @ulist: the ulist to initialize * * Note: don't use this function to init an already used ulist, use * ulist_reinit instead. */ void ulist_init(struct ulist *ulist) { ulist->nnodes = 0; ulist->nodes = ulist->int_nodes; ulist->nodes_alloced = ULIST_SIZE; ulist->root = RB_ROOT; } EXPORT_SYMBOL(ulist_init); /** * ulist_fini - free up additionally allocated memory for the ulist * @ulist: the ulist from which to free the additional memory * * This is useful in cases where the base 'struct ulist' has been statically * allocated. */ void ulist_fini(struct ulist *ulist) { /* * The first ULIST_SIZE elements are stored inline in struct ulist. * Only if more elements are alocated they need to be freed. */ if (ulist->nodes_alloced > ULIST_SIZE) kfree(ulist->nodes); ulist->nodes_alloced = 0; /* in case ulist_fini is called twice */ ulist->root = RB_ROOT; } EXPORT_SYMBOL(ulist_fini); /** * ulist_reinit - prepare a ulist for reuse * @ulist: ulist to be reused * * Free up all additional memory allocated for the list elements and reinit * the ulist. */ void ulist_reinit(struct ulist *ulist) { ulist_fini(ulist); ulist_init(ulist); } EXPORT_SYMBOL(ulist_reinit); /** * ulist_alloc - dynamically allocate a ulist * @gfp_mask: allocation flags to for base allocation * * The allocated ulist will be returned in an initialized state. */ struct ulist *ulist_alloc(gfp_t gfp_mask) { struct ulist *ulist = kmalloc(sizeof(*ulist), gfp_mask); if (!ulist) return NULL; ulist_init(ulist); return ulist; } EXPORT_SYMBOL(ulist_alloc); /** * ulist_free - free dynamically allocated ulist * @ulist: ulist to free * * It is not necessary to call ulist_fini before. */ void ulist_free(struct ulist *ulist) { if (!ulist) return; ulist_fini(ulist); kfree(ulist); } EXPORT_SYMBOL(ulist_free); static struct ulist_node *ulist_rbtree_search(struct ulist *ulist, u64 val) { struct rb_node *n = ulist->root.rb_node; struct ulist_node *u = NULL; while (n) { u = rb_entry(n, struct ulist_node, rb_node); if (u->val < val) n = n->rb_right; else if (u->val > val) n = n->rb_left; else return u; } return NULL; } static int ulist_rbtree_insert(struct ulist *ulist, struct ulist_node *ins) { struct rb_node **p = &ulist->root.rb_node; struct rb_node *parent = NULL; struct ulist_node *cur = NULL; while (*p) { parent = *p; cur = rb_entry(parent, struct ulist_node, rb_node); if (cur->val < ins->val) p = &(*p)->rb_right; else if (cur->val > ins->val) p = &(*p)->rb_left; else return -EEXIST; } rb_link_node(&ins->rb_node, parent, p); rb_insert_color(&ins->rb_node, &ulist->root); return 0; } /** * ulist_add - add an element to the ulist * @ulist: ulist to add the element to * @val: value to add to ulist * @aux: auxiliary value to store along with val * @gfp_mask: flags to use for allocation * * Note: locking must be provided by the caller. In case of rwlocks write * locking is needed * * Add an element to a ulist. The @val will only be added if it doesn't * already exist. If it is added, the auxiliary value @aux is stored along with * it. In case @val already exists in the ulist, @aux is ignored, even if * it differs from the already stored value. * * ulist_add returns 0 if @val already exists in ulist and 1 if @val has been * inserted. * In case of allocation failure -ENOMEM is returned and the ulist stays * unaltered. */ int ulist_add(struct ulist *ulist, u64 val, u64 aux, gfp_t gfp_mask) { return ulist_add_merge(ulist, val, aux, NULL, gfp_mask); } int ulist_add_merge(struct ulist *ulist, u64 val, u64 aux, u64 *old_aux, gfp_t gfp_mask) { int ret = 0; struct ulist_node *node = NULL; node = ulist_rbtree_search(ulist, val); if (node) { if (old_aux) *old_aux = node->aux; return 0; } if (ulist->nnodes >= ulist->nodes_alloced) { u64 new_alloced = ulist->nodes_alloced + 128; struct ulist_node *new_nodes; void *old = NULL; /* * if nodes_alloced == ULIST_SIZE no memory has been allocated * yet, so pass NULL to krealloc */ if (ulist->nodes_alloced > ULIST_SIZE) old = ulist->nodes; new_nodes = krealloc(old, sizeof(*new_nodes) * new_alloced, gfp_mask); if (!new_nodes) return -ENOMEM; if (!old) memcpy(new_nodes, ulist->int_nodes, sizeof(ulist->int_nodes)); ulist->nodes = new_nodes; ulist->nodes_alloced = new_alloced; } ulist->nodes[ulist->nnodes].val = val; ulist->nodes[ulist->nnodes].aux = aux; ret = ulist_rbtree_insert(ulist, &ulist->nodes[ulist->nnodes]); BUG_ON(ret); ++ulist->nnodes; return 1; } EXPORT_SYMBOL(ulist_add); /** * ulist_next - iterate ulist * @ulist: ulist to iterate * @uiter: iterator variable, initialized with ULIST_ITER_INIT(&iterator) * * Note: locking must be provided by the caller. In case of rwlocks only read * locking is needed * * This function is used to iterate an ulist. * It returns the next element from the ulist or %NULL when the * end is reached. No guarantee is made with respect to the order in which * the elements are returned. They might neither be returned in order of * addition nor in ascending order. * It is allowed to call ulist_add during an enumeration. Newly added items * are guaranteed to show up in the running enumeration. */ struct ulist_node *ulist_next(struct ulist *ulist, struct ulist_iterator *uiter) { if (ulist->nnodes == 0) return NULL; if (uiter->i < 0 || uiter->i >= ulist->nnodes) return NULL; return &ulist->nodes[uiter->i++]; } EXPORT_SYMBOL(ulist_next);