diff options
Diffstat (limited to 'testsuite/epoch01/test_main.c')
-rw-r--r-- | testsuite/epoch01/test_main.c | 596 |
1 files changed, 596 insertions, 0 deletions
diff --git a/testsuite/epoch01/test_main.c b/testsuite/epoch01/test_main.c new file mode 100644 index 00000000..7f8003b5 --- /dev/null +++ b/testsuite/epoch01/test_main.c @@ -0,0 +1,596 @@ +/* + * Copyright (c) 2018 embedded brains GmbH. All rights reserved. + * + * embedded brains GmbH + * Dornierstr. 4 + * 82178 Puchheim + * Germany + * <rtems@embedded-brains.de> + * + * Redistribution and use in source and binary forms, with or without + * modification, are permitted provided that the following conditions + * are met: + * 1. Redistributions of source code must retain the above copyright + * notice, this list of conditions and the following disclaimer. + * 2. Redistributions in binary form must reproduce the above copyright + * notice, this list of conditions and the following disclaimer in the + * documentation and/or other materials provided with the distribution. + * + * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND + * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE + * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE + * ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE + * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL + * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS + * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) + * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT + * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY + * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF + * SUCH DAMAGE. + */ + +#include <machine/rtems-bsd-kernel-space.h> + +#include <sys/param.h> +#include <sys/types.h> +#include <sys/systm.h> +#include <sys/lock.h> +#include <sys/mutex.h> +#include <sys/epoch.h> + +#include <assert.h> +#include <inttypes.h> +#include <stdio.h> +#include <stdlib.h> +#include <string.h> + +#include <ck_queue.h> + +#include <rtems.h> +#include <rtems/cpuuse.h> +#include <rtems/test.h> +#include <rtems/thread.h> + +#define TEST_NAME "LIBBSD EPOCH 1" + +#define TEST_XML_NAME "TestEpoch01" + +#define CPU_COUNT 32 + +typedef struct { + uint32_t counter[CPU_COUNT]; + uint32_t removals[CPU_COUNT]; + uint32_t item_counter[CPU_COUNT][CPU_COUNT]; +} test_stats; + +typedef struct test_item { + CK_SLIST_ENTRY(test_item) link; + struct epoch_context ec; + size_t index; + size_t worker_index; +} test_item; + +typedef struct { + rtems_test_parallel_context base; + size_t active_workers; + CK_SLIST_HEAD(, test_item) item_list; + rtems_mutex mtx[2]; + test_item items[CPU_COUNT]; + test_stats stats; +} test_context; + +static test_context test_instance; + +static rtems_interval +test_duration(void) +{ + + return (1 * rtems_clock_get_ticks_per_second()); +} + +static rtems_interval +test_init(rtems_test_parallel_context *base, void *arg, size_t active_workers) +{ + test_context *ctx; + + ctx = (test_context *)base; + memset(&ctx->stats, 0, sizeof(ctx->stats)); + return (test_duration()); +} + +static void +test_fini(rtems_test_parallel_context *base, const char *name, + size_t active_workers) +{ + test_context *ctx; + size_t i; + + ctx = (test_context *)base; + + printf(" <%s activeWorker=\"%zu\">\n", name, active_workers); + + for (i = 0; i < active_workers; ++i) { + printf(" <Counter worker=\"%zu\">%" PRIu32 "</Counter>\n", + i, ctx->stats.counter[i]); + } + + for (i = 0; i < active_workers; ++i) { + uint32_t r; + + r = ctx->stats.removals[i]; + if (r > 0) { + printf(" <Removals worker=\"%zu\">%" PRIu32 + "</Removals>\n", i, r); + } + } + + printf(" </%s>\n", name); +} + +static void +test_enter_exit_body(rtems_test_parallel_context *base, void *arg, + size_t active_workers, size_t worker_index) +{ + test_context *ctx = (test_context *)base; + epoch_t e = global_epoch; + uint32_t counter; + + ctx = (test_context *)base; + e = global_epoch; + counter = 0; + + while (!rtems_test_parallel_stop_job(&ctx->base)) { + epoch_enter(e); + ++counter; + epoch_exit(e); + } + + ctx->stats.counter[worker_index] = counter; +} + +static void +test_enter_exit_fini(rtems_test_parallel_context *base, void *arg, + size_t active_workers) +{ + + test_fini(base, "EnterExit", active_workers); +} + +static rtems_interval +test_list_init(rtems_test_parallel_context *base, void *arg, + size_t active_workers) +{ + test_context *ctx; + size_t i; + + ctx = (test_context *)base; + ctx->active_workers = active_workers; + CK_SLIST_INIT(&ctx->item_list); + + for (i = 0; i < active_workers; ++i) { + test_item *item = &ctx->items[i]; + + CK_SLIST_INSERT_HEAD(&ctx->item_list, item, link); + item->index = i; + + if (i == 0) { + item->worker_index = 0; + } else { + item->worker_index = CPU_COUNT; + } + } + + return (test_init(&ctx->base, arg, active_workers)); +} + +static void +test_list_callback(epoch_context_t ec) +{ + test_context *ctx; + test_item *item; + test_item *next; + + ctx = &test_instance; + item = __containerof(ec, struct test_item, ec); + item->worker_index = (item->worker_index + 1) % ctx->active_workers; + next = CK_SLIST_NEXT(item, link); + + if (next != NULL) { + CK_SLIST_INSERT_AFTER(next, item, link); + } else { + CK_SLIST_INSERT_HEAD(&ctx->item_list, item, link); + } +} + +static void +test_enter_list_op_exit_body(rtems_test_parallel_context *base, void *arg, + size_t active_workers, size_t worker_index) +{ + test_context *ctx; + epoch_t e; + uint32_t counter; + uint32_t removals; + uint32_t item_counter[CPU_COUNT]; + + ctx = (test_context *)base; + e = global_epoch; + counter = 0; + removals = 0; + memset(item_counter, 0, sizeof(item_counter)); + + while (!rtems_test_parallel_stop_job(&ctx->base)) { + test_item *prev; + test_item *item; + test_item *tmp; + test_item *rm; + + epoch_enter(e); + ++counter; + + prev = NULL; + rm = NULL; + CK_SLIST_FOREACH_SAFE(item, &ctx->item_list, link, tmp) { + ++item_counter[item->index]; + + if (item->worker_index == worker_index) { + ++removals; + rm = item; + + if (prev != NULL) { + CK_SLIST_REMOVE_AFTER(prev, link); + } else { + CK_SLIST_REMOVE_HEAD(&ctx->item_list, + link); + } + } + + prev = item; + } + + epoch_exit(e); + + if (rm != NULL) { + epoch_call(e, &rm->ec, test_list_callback); + epoch_wait(e); + } + } + + ctx->stats.counter[worker_index] = counter; + ctx->stats.removals[worker_index] = removals; + memcpy(ctx->stats.item_counter[worker_index], item_counter, + sizeof(ctx->stats.item_counter[worker_index])); +} + +static void +test_enter_list_op_exit_fini(rtems_test_parallel_context *base, void *arg, + size_t active_workers) +{ + + test_fini(base, "EnterListOpExit", active_workers); +} + +static void +test_enter_exit_preempt_body(rtems_test_parallel_context *base, void *arg, + size_t active_workers, size_t worker_index) +{ + test_context *ctx = (test_context *)base; + epoch_t e = global_epoch; + uint32_t counter; + + ctx = (test_context *)base; + e = global_epoch; + counter = 0; + + while (!rtems_test_parallel_stop_job(&ctx->base)) { + epoch_enter_preempt(e); + ++counter; + epoch_exit_preempt(e); + } + + ctx->stats.counter[worker_index] = counter; +} + +static void +test_enter_exit_preempt_fini(rtems_test_parallel_context *base, + void *arg, size_t active_workers) +{ + + test_fini(base, "EnterExitPreempt", active_workers); +} + +static void +test_enter_list_op_exit_preempt_body(rtems_test_parallel_context *base, + void *arg, size_t active_workers, size_t worker_index) +{ + test_context *ctx; + epoch_t e; + uint32_t counter; + uint32_t removals; + uint32_t item_counter[CPU_COUNT]; + + ctx = (test_context *)base; + e = global_epoch; + counter = 0; + removals = 0; + memset(item_counter, 0, sizeof(item_counter)); + + while (!rtems_test_parallel_stop_job(&ctx->base)) { + test_item *prev; + test_item *item; + test_item *tmp; + test_item *rm; + + epoch_enter_preempt(e); + ++counter; + + prev = NULL; + rm = NULL; + CK_SLIST_FOREACH_SAFE(item, &ctx->item_list, link, tmp) { + ++item_counter[item->index]; + + if (item->worker_index == worker_index) { + ++removals; + rm = item; + + if (prev != NULL) { + CK_SLIST_REMOVE_AFTER(prev, link); + } else { + CK_SLIST_REMOVE_HEAD(&ctx->item_list, + link); + } + } + + prev = item; + } + + epoch_exit_preempt(e); + + if (rm != NULL) { + epoch_call(e, &rm->ec, test_list_callback); + epoch_wait_preempt(e); + } + } + + ctx->stats.counter[worker_index] = counter; + ctx->stats.removals[worker_index] = removals; + memcpy(ctx->stats.item_counter[worker_index], item_counter, + sizeof(ctx->stats.item_counter[worker_index])); +} + +static void +test_enter_list_op_exit_preempt_fini(rtems_test_parallel_context *base, + void *arg, size_t active_workers) +{ + + test_fini(base, "EnterListOpExitPreempt", active_workers); +} + +static void +test_thread_local_mutex_body(rtems_test_parallel_context *base, void *arg, + size_t active_workers, size_t worker_index) +{ + test_context *ctx = (test_context *)base; + rtems_mutex mtx; + uint32_t counter; + + ctx = (test_context *)base; + rtems_mutex_init(&mtx, "test"); + counter = 0; + + while (!rtems_test_parallel_stop_job(&ctx->base)) { + rtems_mutex_lock(&mtx); + ++counter; + rtems_mutex_unlock(&mtx); + } + + rtems_mutex_destroy(&mtx); + ctx->stats.counter[worker_index] = counter; +} + +static void +test_thread_local_mutex_fini(rtems_test_parallel_context *base, + void *arg, size_t active_workers) +{ + + test_fini(base, "ThreadLocalMutex", active_workers); +} + +static void +test_enter_mutex_exit_preempt_body(rtems_test_parallel_context *base, + void *arg, size_t active_workers, size_t worker_index) +{ + test_context *ctx = (test_context *)base; + epoch_t e = global_epoch; + uint32_t counter; + rtems_mutex *mtx; + + ctx = (test_context *)base; + e = global_epoch; + counter = 0; + mtx = &ctx->mtx[worker_index % RTEMS_ARRAY_SIZE(ctx->mtx)]; + + while (!rtems_test_parallel_stop_job(&ctx->base)) { + epoch_enter_preempt(e); + rtems_mutex_lock(mtx); + ++counter; + rtems_mutex_unlock(mtx); + epoch_exit_preempt(e); + } + + ctx->stats.counter[worker_index] = counter; +} + +static void +test_enter_mutex_exit_preempt_fini(rtems_test_parallel_context *base, + void *arg, size_t active_workers) +{ + + test_fini(base, "EnterMutexExitPreempt", active_workers); +} + +static const rtems_test_parallel_job test_jobs[] = { + { + .init = test_init, + .body = test_enter_exit_body, + .fini = test_enter_exit_fini, + .cascade = true + }, { + .init = test_list_init, + .body = test_enter_list_op_exit_body, + .fini = test_enter_list_op_exit_fini, + .cascade = true + }, { + .init = test_init, + .body = test_enter_exit_preempt_body, + .fini = test_enter_exit_preempt_fini, + .cascade = true + }, { + .init = test_list_init, + .body = test_enter_list_op_exit_preempt_body, + .fini = test_enter_list_op_exit_preempt_fini, + .cascade = true + }, { + .init = test_init, + .body = test_thread_local_mutex_body, + .fini = test_thread_local_mutex_fini, + .cascade = true + }, { + .init = test_init, + .body = test_enter_mutex_exit_preempt_body, + .fini = test_enter_mutex_exit_preempt_fini, + .cascade = true + } +}; + +static void +set_affinity(rtems_id task, size_t cpu_index) +{ + rtems_status_code sc; + cpu_set_t set; + rtems_id sched; + rtems_task_priority prio; + + sc = rtems_scheduler_ident_by_processor(cpu_index, &sched); + assert(sc == RTEMS_SUCCESSFUL); + + sc = rtems_task_set_priority(task, RTEMS_CURRENT_PRIORITY, &prio); + assert(sc == RTEMS_SUCCESSFUL); + + sc = rtems_task_set_scheduler(task, sched, prio); + assert(sc == RTEMS_SUCCESSFUL); + + CPU_ZERO(&set); + CPU_SET((int)cpu_index, &set); + sc = rtems_task_set_affinity(task, sizeof(set), &set); + assert(sc == RTEMS_SUCCESSFUL); +} + +static void +setup_worker(rtems_test_parallel_context *base, size_t worker_index, + rtems_id worker_id) +{ + + set_affinity(worker_id, worker_index); +} + +static void +test_main(void) +{ + test_context *ctx; + + ctx = &test_instance; + rtems_mutex_init(&ctx->mtx[0], "test 0"); + rtems_mutex_init(&ctx->mtx[1], "test 1"); + + printf("<" TEST_XML_NAME ">\n"); + + setup_worker(&ctx->base, 0, rtems_task_self()); + rtems_test_parallel(&ctx->base, setup_worker, &test_jobs[0], + RTEMS_ARRAY_SIZE(test_jobs)); + + printf("</" TEST_XML_NAME ">\n"); + rtems_mutex_destroy(&ctx->mtx[0]); + rtems_mutex_destroy(&ctx->mtx[1]); + exit(0); +} + +#define CONFIGURE_MAXIMUM_PROCESSORS CPU_COUNT + +#ifdef RTEMS_SMP + +#define CONFIGURE_SCHEDULER_EDF_SMP + +#include <rtems/scheduler.h> + +RTEMS_SCHEDULER_EDF_SMP(a, CPU_COUNT); +RTEMS_SCHEDULER_EDF_SMP(b, CPU_COUNT); +RTEMS_SCHEDULER_EDF_SMP(c, CPU_COUNT); +RTEMS_SCHEDULER_EDF_SMP(d, CPU_COUNT); +RTEMS_SCHEDULER_EDF_SMP(e, CPU_COUNT); +RTEMS_SCHEDULER_EDF_SMP(g, CPU_COUNT); +RTEMS_SCHEDULER_EDF_SMP(h, CPU_COUNT); +RTEMS_SCHEDULER_EDF_SMP(i, CPU_COUNT); +RTEMS_SCHEDULER_EDF_SMP(j, CPU_COUNT); +RTEMS_SCHEDULER_EDF_SMP(k, CPU_COUNT); +RTEMS_SCHEDULER_EDF_SMP(l, CPU_COUNT); +RTEMS_SCHEDULER_EDF_SMP(m, CPU_COUNT); +RTEMS_SCHEDULER_EDF_SMP(n, CPU_COUNT); +RTEMS_SCHEDULER_EDF_SMP(o, CPU_COUNT); +RTEMS_SCHEDULER_EDF_SMP(p, CPU_COUNT); +RTEMS_SCHEDULER_EDF_SMP(q, CPU_COUNT); + +#define CONFIGURE_SCHEDULER_TABLE_ENTRIES \ + RTEMS_SCHEDULER_TABLE_EDF_SMP(a, rtems_build_name(' ', ' ', ' ', 'a')), \ + RTEMS_SCHEDULER_TABLE_EDF_SMP(b, rtems_build_name(' ', ' ', ' ', 'b')), \ + RTEMS_SCHEDULER_TABLE_EDF_SMP(c, rtems_build_name(' ', ' ', ' ', 'c')), \ + RTEMS_SCHEDULER_TABLE_EDF_SMP(d, rtems_build_name(' ', ' ', ' ', 'd')), \ + RTEMS_SCHEDULER_TABLE_EDF_SMP(e, rtems_build_name(' ', ' ', ' ', 'e')), \ + RTEMS_SCHEDULER_TABLE_EDF_SMP(g, rtems_build_name(' ', ' ', ' ', 'g')), \ + RTEMS_SCHEDULER_TABLE_EDF_SMP(h, rtems_build_name(' ', ' ', ' ', 'h')), \ + RTEMS_SCHEDULER_TABLE_EDF_SMP(i, rtems_build_name(' ', ' ', ' ', 'i')), \ + RTEMS_SCHEDULER_TABLE_EDF_SMP(j, rtems_build_name(' ', ' ', ' ', 'j')), \ + RTEMS_SCHEDULER_TABLE_EDF_SMP(k, rtems_build_name(' ', ' ', ' ', 'k')), \ + RTEMS_SCHEDULER_TABLE_EDF_SMP(l, rtems_build_name(' ', ' ', ' ', 'l')), \ + RTEMS_SCHEDULER_TABLE_EDF_SMP(m, rtems_build_name(' ', ' ', ' ', 'm')), \ + RTEMS_SCHEDULER_TABLE_EDF_SMP(n, rtems_build_name(' ', ' ', ' ', 'n')), \ + RTEMS_SCHEDULER_TABLE_EDF_SMP(o, rtems_build_name(' ', ' ', ' ', 'o')), \ + RTEMS_SCHEDULER_TABLE_EDF_SMP(p, rtems_build_name(' ', ' ', ' ', 'p')), \ + RTEMS_SCHEDULER_TABLE_EDF_SMP(q, rtems_build_name(' ', ' ', ' ', 'q')) \ + +#define CONFIGURE_SCHEDULER_ASSIGNMENTS \ + RTEMS_SCHEDULER_ASSIGN(0, RTEMS_SCHEDULER_ASSIGN_PROCESSOR_MANDATORY), \ + RTEMS_SCHEDULER_ASSIGN(0, RTEMS_SCHEDULER_ASSIGN_PROCESSOR_OPTIONAL), \ + RTEMS_SCHEDULER_ASSIGN(1, RTEMS_SCHEDULER_ASSIGN_PROCESSOR_OPTIONAL), \ + RTEMS_SCHEDULER_ASSIGN(1, RTEMS_SCHEDULER_ASSIGN_PROCESSOR_OPTIONAL), \ + RTEMS_SCHEDULER_ASSIGN(2, RTEMS_SCHEDULER_ASSIGN_PROCESSOR_OPTIONAL), \ + RTEMS_SCHEDULER_ASSIGN(2, RTEMS_SCHEDULER_ASSIGN_PROCESSOR_OPTIONAL), \ + RTEMS_SCHEDULER_ASSIGN(3, RTEMS_SCHEDULER_ASSIGN_PROCESSOR_OPTIONAL), \ + RTEMS_SCHEDULER_ASSIGN(3, RTEMS_SCHEDULER_ASSIGN_PROCESSOR_OPTIONAL), \ + RTEMS_SCHEDULER_ASSIGN(4, RTEMS_SCHEDULER_ASSIGN_PROCESSOR_OPTIONAL), \ + RTEMS_SCHEDULER_ASSIGN(4, RTEMS_SCHEDULER_ASSIGN_PROCESSOR_OPTIONAL), \ + RTEMS_SCHEDULER_ASSIGN(5, RTEMS_SCHEDULER_ASSIGN_PROCESSOR_OPTIONAL), \ + RTEMS_SCHEDULER_ASSIGN(5, RTEMS_SCHEDULER_ASSIGN_PROCESSOR_OPTIONAL), \ + RTEMS_SCHEDULER_ASSIGN(6, RTEMS_SCHEDULER_ASSIGN_PROCESSOR_OPTIONAL), \ + RTEMS_SCHEDULER_ASSIGN(6, RTEMS_SCHEDULER_ASSIGN_PROCESSOR_OPTIONAL), \ + RTEMS_SCHEDULER_ASSIGN(7, RTEMS_SCHEDULER_ASSIGN_PROCESSOR_OPTIONAL), \ + RTEMS_SCHEDULER_ASSIGN(7, RTEMS_SCHEDULER_ASSIGN_PROCESSOR_OPTIONAL), \ + RTEMS_SCHEDULER_ASSIGN(8, RTEMS_SCHEDULER_ASSIGN_PROCESSOR_OPTIONAL), \ + RTEMS_SCHEDULER_ASSIGN(8, RTEMS_SCHEDULER_ASSIGN_PROCESSOR_OPTIONAL), \ + RTEMS_SCHEDULER_ASSIGN(9, RTEMS_SCHEDULER_ASSIGN_PROCESSOR_OPTIONAL), \ + RTEMS_SCHEDULER_ASSIGN(9, RTEMS_SCHEDULER_ASSIGN_PROCESSOR_OPTIONAL), \ + RTEMS_SCHEDULER_ASSIGN(10, RTEMS_SCHEDULER_ASSIGN_PROCESSOR_OPTIONAL), \ + RTEMS_SCHEDULER_ASSIGN(10, RTEMS_SCHEDULER_ASSIGN_PROCESSOR_OPTIONAL), \ + RTEMS_SCHEDULER_ASSIGN(11, RTEMS_SCHEDULER_ASSIGN_PROCESSOR_OPTIONAL), \ + RTEMS_SCHEDULER_ASSIGN(11, RTEMS_SCHEDULER_ASSIGN_PROCESSOR_OPTIONAL), \ + RTEMS_SCHEDULER_ASSIGN(12, RTEMS_SCHEDULER_ASSIGN_PROCESSOR_OPTIONAL), \ + RTEMS_SCHEDULER_ASSIGN(12, RTEMS_SCHEDULER_ASSIGN_PROCESSOR_OPTIONAL), \ + RTEMS_SCHEDULER_ASSIGN(13, RTEMS_SCHEDULER_ASSIGN_PROCESSOR_OPTIONAL), \ + RTEMS_SCHEDULER_ASSIGN(13, RTEMS_SCHEDULER_ASSIGN_PROCESSOR_OPTIONAL), \ + RTEMS_SCHEDULER_ASSIGN(14, RTEMS_SCHEDULER_ASSIGN_PROCESSOR_OPTIONAL), \ + RTEMS_SCHEDULER_ASSIGN(14, RTEMS_SCHEDULER_ASSIGN_PROCESSOR_OPTIONAL), \ + RTEMS_SCHEDULER_ASSIGN(15, RTEMS_SCHEDULER_ASSIGN_PROCESSOR_OPTIONAL), \ + RTEMS_SCHEDULER_ASSIGN(15, RTEMS_SCHEDULER_ASSIGN_PROCESSOR_OPTIONAL) + +#endif /* RTEMS_SMP */ + +#include <rtems/bsd/test/default-init.h> |