/* * UCW Library -- Optimized Array Sorter * * (c) 2003--2007 Martin Mares * * This software may be freely distributed and used according to the terms * of the GNU Lesser General Public License. */ #undef LOCAL_DEBUG #include #include #include #include #define ASORT_MIN_SHIFT 2 #define ASORT_TRACE(x...) ASORT_XTRACE(1, x) #define ASORT_XTRACE(level, x...) do { if (sorter_trace_array >= level) msg(L_DEBUG, x); } while(0) static void asort_radix(struct asort_context *ctx, void *array, void *buffer, uint num_elts, uint hash_bits, uint swapped_output) { // swap_output == 0 if result should be returned in `array', otherwise in `buffer' uint buckets = (1 << ctx->radix_bits); uint shift = (hash_bits > ctx->radix_bits) ? (hash_bits - ctx->radix_bits) : 0; uint cnt[buckets]; #if 0 static int reported[64]; if (!reported[hash_bits]++) #endif DBG(">>> n=%u h=%d s=%d sw=%d", num_elts, hash_bits, shift, swapped_output); bzero(cnt, sizeof(cnt)); ctx->radix_count(array, num_elts, cnt, shift); uint pos = 0; for (uint i=0; iradix_split(array, buffer, num_elts, cnt, shift); pos = 0; for (uint i=0; iradix_threshold || shift < ASORT_MIN_SHIFT) { ctx->quicksort(buffer, n); if (!swapped_output) memcpy(array, buffer, n * ctx->elt_size); } else asort_radix(ctx, buffer, array, n, shift, !swapped_output); array += n * ctx->elt_size; buffer += n * ctx->elt_size; pos = cnt[i]; } } #ifdef CONFIG_UCW_THREADS #include #include #include static uint asort_threads_use_count; static uint asort_threads_ready; static struct worker_pool asort_thread_pool; static uint rs_estimate_stack(void) { // Stack space needed by the recursive radix-sorter uint ctrsize = sizeof(uint) * (1 << CONFIG_UCW_RADIX_SORTER_BITS); uint maxdepth = (64 / CONFIG_UCW_RADIX_SORTER_BITS) + 1; return ctrsize * maxdepth; } void asort_start_threads(uint run) { ucwlib_lock(); asort_threads_use_count++; if (run && !asort_threads_ready) { // XXX: If somebody overrides the radix-sorter parameters to insane values, // he also should override the stack size to insane values. asort_thread_pool.stack_size = ucwlib_thread_stack_size + rs_estimate_stack(); asort_thread_pool.num_threads = sorter_threads; ASORT_TRACE("Initializing thread pool (%d threads, %dK stack)", sorter_threads, asort_thread_pool.stack_size >> 10); worker_pool_init(&asort_thread_pool); asort_threads_ready = 1; } ucwlib_unlock(); } void asort_stop_threads(void) { ucwlib_lock(); if (!--asort_threads_use_count && asort_threads_ready) { ASORT_TRACE("Shutting down thread pool"); worker_pool_cleanup(&asort_thread_pool); asort_threads_ready = 0; } ucwlib_unlock(); } struct qs_work { struct work w; struct asort_context *ctx; void *array; uint num_elts; int left, right; #define LR_UNDEF -100 }; static void qs_handle_work(struct worker_thread *thr UNUSED, struct work *ww) { struct qs_work *w = (struct qs_work *) ww; struct asort_context *ctx = w->ctx; DBG("Thread %d: got %u elts", thr->id, w->num_elts); if (w->num_elts < ctx->thread_threshold) { ctx->quicksort(w->array, w->num_elts); w->left = w->right = LR_UNDEF; } else ctx->quicksplit(w->array, w->num_elts, &w->left, &w->right); DBG("Thread %d: returning l=%u r=%u", thr->id, w->left, w->right); } static struct qs_work * qs_alloc_work(struct asort_context *ctx) { struct qs_work *w = ep_alloc(ctx->eltpool); w->w.priority = 0; w->w.go = qs_handle_work; w->ctx = ctx; return w; } static void threaded_quicksort(struct asort_context *ctx) { struct work_queue q; struct qs_work *v, *w; asort_start_threads(1); work_queue_init(&asort_thread_pool, &q); ctx->eltpool = ep_new(sizeof(struct qs_work), 1000); w = qs_alloc_work(ctx); w->array = ctx->array; w->num_elts = ctx->num_elts; work_submit(&q, &w->w); while (v = (struct qs_work *) work_wait(&q)) { if (v->left != LR_UNDEF) { if (v->right > 0) { w = qs_alloc_work(ctx); w->array = v->array; w->num_elts = v->right + 1; w->w.priority = v->w.priority + 1; work_submit(&q, &w->w); } if (v->left < (int)v->num_elts - 1) { w = qs_alloc_work(ctx); w->array = v->array + v->left * ctx->elt_size; w->num_elts = v->num_elts - v->left; w->w.priority = v->w.priority + 1; work_submit(&q, &w->w); } } ep_free(ctx->eltpool, v); } ep_delete(ctx->eltpool); work_queue_cleanup(&q); asort_stop_threads(); } struct rs_work { struct work w; struct asort_context *ctx; void *array, *buffer; // Like asort_radix(). uint num_elts; uint shift; uint swap_output; uint cnt[0]; }; static void rs_count(struct worker_thread *thr UNUSED, struct work *ww) { struct rs_work *w = (struct rs_work *) ww; DBG("Thread %d: Counting %u items, shift=%d", thr->id, w->num_elts, w->shift); w->ctx->radix_count(w->array, w->num_elts, w->cnt, w->shift); DBG("Thread %d: Counting done", thr->id); } static void rs_split(struct worker_thread *thr UNUSED, struct work *ww) { struct rs_work *w = (struct rs_work *) ww; DBG("Thread %d: Splitting %u items, shift=%d", thr->id, w->num_elts, w->shift); w->ctx->radix_split(w->array, w->buffer, w->num_elts, w->cnt, w->shift); DBG("Thread %d: Splitting done", thr->id); } static void rs_finish(struct worker_thread *thr UNUSED, struct work *ww) { struct rs_work *w = (struct rs_work *) ww; if (thr) DBG("Thread %d: Finishing %u items, shift=%d", thr->id, w->num_elts, w->shift); if (w->shift < ASORT_MIN_SHIFT || w->num_elts < w->ctx->radix_threshold) { w->ctx->quicksort(w->array, w->num_elts); if (w->swap_output) memcpy(w->buffer, w->array, w->num_elts * w->ctx->elt_size); } else asort_radix(w->ctx, w->array, w->buffer, w->num_elts, w->shift, w->swap_output); if (thr) DBG("Thread %d: Finishing done", thr->id); } static void rs_wait_small(struct asort_context *ctx) { struct rs_work *w; while (w = (struct rs_work *) work_wait(ctx->rs_work_queue)) { DBG("Reaping small chunk of %u items", w->num_elts); ep_free(ctx->eltpool, w); } } static void rs_radix(struct asort_context *ctx, void *array, void *buffer, uint num_elts, uint hash_bits, uint swapped_output) { uint buckets = (1 << ctx->radix_bits); uint shift = (hash_bits > ctx->radix_bits) ? (hash_bits - ctx->radix_bits) : 0; uint cnt[buckets]; uint blksize = num_elts / sorter_threads; DBG(">>> n=%u h=%d s=%d blk=%u sw=%d", num_elts, hash_bits, shift, blksize, swapped_output); // If there are any small chunks in progress, wait for them to finish rs_wait_small(ctx); // Start parallel counting void *iptr = array; for (uint i=0; irs_works[i]; w->w.priority = 0; w->w.go = rs_count; w->ctx = ctx; w->array = iptr; w->buffer = buffer; w->num_elts = blksize; if (i == sorter_threads-1) w->num_elts += num_elts % sorter_threads; w->shift = shift; iptr += w->num_elts * ctx->elt_size; bzero(w->cnt, sizeof(uint) * buckets); work_submit(ctx->rs_work_queue, &w->w); } // Get bucket sizes from the counts bzero(cnt, sizeof(cnt)); for (uint i=0; irs_work_queue); ASSERT(w); for (uint j=0; jcnt[j]; } // Calculate bucket starts uint pos = 0; for (uint i=0; irs_works[i]; w->w.go = rs_split; for (uint j=0; jcnt[j]; w->cnt[j] = cnt[j]; cnt[j] += k; } work_submit(ctx->rs_work_queue, &w->w); } ASSERT(cnt[buckets-1] == num_elts); // Wait for splits to finish while (work_wait(ctx->rs_work_queue)) ; // Recurse on buckets pos = 0; for (uint i=0; ithread_threshold || shift < ASORT_MIN_SHIFT) { struct rs_work *w = ep_alloc(ctx->eltpool); w->w.priority = 0; w->w.go = rs_finish; w->ctx = ctx; w->array = buffer; w->buffer = array; w->num_elts = n; w->shift = shift; w->swap_output = !swapped_output; if (n < ctx->thread_chunk) { DBG("Sorting block %u+%u inline", pos, n); rs_finish(NULL, &w->w); ep_free(ctx->eltpool, w); } else { DBG("Scheduling block %u+%u", pos, n); work_submit(ctx->rs_work_queue, &w->w); } } else rs_radix(ctx, buffer, array, n, shift, !swapped_output); pos = cnt[i]; array += n * ctx->elt_size; buffer += n * ctx->elt_size; } } static void threaded_radixsort(struct asort_context *ctx, uint swap) { struct work_queue q; asort_start_threads(1); work_queue_init(&asort_thread_pool, &q); // Prepare work structures for counting and splitting. // We use big_alloc(), because we want to avoid cacheline aliasing between threads. ctx->rs_work_queue = &q; ctx->rs_works = alloca(sizeof(struct rs_work *) * sorter_threads); for (uint i=0; irs_works[i] = big_alloc(sizeof(struct rs_work) + sizeof(uint) * (1 << ctx->radix_bits)); // Prepare a pool for all remaining small bits which will be sorted on background. ctx->eltpool = ep_new(sizeof(struct rs_work), 1000); // Do the big splitting rs_radix(ctx, ctx->array, ctx->buffer, ctx->num_elts, ctx->hash_bits, swap); for (uint i=0; irs_works[i], sizeof(struct rs_work) + sizeof(uint) * (1 << ctx->radix_bits)); // Finish the small blocks rs_wait_small(ctx); ASSERT(!ctx->eltpool->num_allocated); ep_delete(ctx->eltpool); work_queue_cleanup(&q); asort_stop_threads(); } #else void asort_start_threads(uint run UNUSED) { } void asort_stop_threads(void) { } #endif static uint predict_swap(struct asort_context *ctx) { uint bits = ctx->radix_bits; uint elts = ctx->num_elts; uint swap = 0; while (elts >= ctx->radix_threshold && bits >= ASORT_MIN_SHIFT) { DBG("Predicting pass: %u elts, %d bits", elts, bits); swap = !swap; elts >>= ctx->radix_bits; bits = MAX(bits, ctx->radix_bits) - ctx->radix_bits; } return swap; } void asort_run(struct asort_context *ctx) { ctx->thread_threshold = MIN(sorter_thread_threshold / ctx->elt_size, ~0U); ctx->thread_chunk = MIN(sorter_thread_chunk / ctx->elt_size, ~0U); ctx->radix_threshold = MIN(sorter_radix_threshold / ctx->elt_size, ~0U); ASORT_TRACE("Array-sorting %u items per %u bytes, hash_bits=%d", ctx->num_elts, ctx->elt_size, ctx->hash_bits); ASORT_XTRACE(2, "Limits: thread_threshold=%u, thread_chunk=%u, radix_threshold=%u", ctx->thread_threshold, ctx->thread_chunk, ctx->radix_threshold); uint allow_threads UNUSED = (sorter_threads > 1 && ctx->num_elts >= ctx->thread_threshold && !(sorter_debug & SORT_DEBUG_ASORT_NO_THREADS)); if (ctx->num_elts < ctx->radix_threshold || ctx->hash_bits <= ASORT_MIN_SHIFT || !ctx->radix_split || (sorter_debug & SORT_DEBUG_ASORT_NO_RADIX)) { #ifdef CONFIG_UCW_THREADS if (allow_threads) { ASORT_XTRACE(2, "Decided to use parallel quicksort"); threaded_quicksort(ctx); } else #endif { ASORT_XTRACE(2, "Decided to use sequential quicksort"); ctx->quicksort(ctx->array, ctx->num_elts); } } else { uint swap = predict_swap(ctx); #ifdef CONFIG_UCW_THREADS if (allow_threads) { ASORT_XTRACE(2, "Decided to use parallel radix-sort (swap=%d)", swap); threaded_radixsort(ctx, swap); } else #endif { ASORT_XTRACE(2, "Decided to use sequential radix-sort (swap=%d)", swap); asort_radix(ctx, ctx->array, ctx->buffer, ctx->num_elts, ctx->hash_bits, swap); } if (swap) ctx->array = ctx->buffer; } ASORT_XTRACE(2, "Array-sort finished"); }