diff options
Diffstat (limited to 'testsuites/validation/tr-tq-flush-priority.c')
-rw-r--r-- | testsuites/validation/tr-tq-flush-priority.c | 424 |
1 files changed, 424 insertions, 0 deletions
diff --git a/testsuites/validation/tr-tq-flush-priority.c b/testsuites/validation/tr-tq-flush-priority.c new file mode 100644 index 0000000000..b8a06c2340 --- /dev/null +++ b/testsuites/validation/tr-tq-flush-priority.c @@ -0,0 +1,424 @@ +/* SPDX-License-Identifier: BSD-2-Clause */ + +/** + * @file + * + * @ingroup ScoreTqReqFlushPriority + */ + +/* + * Copyright (C) 2021 embedded brains GmbH & Co. KG + * + * Redistribution and use in source and binary forms, with or without + * modification, are permitted provided that the following conditions + * are met: + * 1. Redistributions of source code must retain the above copyright + * notice, this list of conditions and the following disclaimer. + * 2. Redistributions in binary form must reproduce the above copyright + * notice, this list of conditions and the following disclaimer in the + * documentation and/or other materials provided with the distribution. + * + * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" + * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE + * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE + * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE + * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR + * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF + * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS + * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN + * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) + * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE + * POSSIBILITY OF SUCH DAMAGE. + */ + +/* + * This file is part of the RTEMS quality process and was automatically + * generated. If you find something that needs to be fixed or + * worded better please post a report or patch to an RTEMS mailing list + * or raise a bug report: + * + * https://www.rtems.org/bugs.html + * + * For information on updating and regenerating please refer to the How-To + * section in the Software Requirements Engineering chapter of the + * RTEMS Software Engineering manual. The manual is provided as a part of + * a release. For development sources please refer to the online + * documentation at: + * + * https://docs.rtems.org + */ + +#ifdef HAVE_CONFIG_H +#include "config.h" +#endif + +#include "tr-tq-flush-priority.h" +#include "tx-support.h" + +#include <rtems/test.h> + +/** + * @defgroup ScoreTqReqFlushPriority spec:/score/tq/req/flush-priority + * + * @ingroup TestsuitesValidationNoClock0 + * + * @{ + */ + +typedef struct { + uint8_t Skip : 1; + uint8_t Pre_Queue_NA : 1; + uint8_t Post_Operation : 2; +} ScoreTqReqFlushPriority_Entry; + +/** + * @brief Test context for spec:/score/tq/req/flush-priority test case. + */ +typedef struct { + /** + * @brief This member contains the call within ISR request. + */ + CallWithinISRRequest request; + + /** + * @brief This member contains a copy of the corresponding + * ScoreTqReqFlushPriority_Run() parameter. + */ + TQContext *tq_ctx; + + /** + * @brief This member contains a copy of the corresponding + * ScoreTqReqFlushPriority_Run() parameter. + */ + bool supports_multiple_priority_queues; + + struct { + /** + * @brief This member defines the pre-condition states for the next action. + */ + size_t pcs[ 1 ]; + + /** + * @brief If this member is true, then the test action loop is executed. + */ + bool in_action_loop; + + /** + * @brief This member contains the next transition map index. + */ + size_t index; + + /** + * @brief This member contains the current transition map entry. + */ + ScoreTqReqFlushPriority_Entry entry; + + /** + * @brief If this member is true, then the current transition variant + * should be skipped. + */ + bool skip; + } Map; +} ScoreTqReqFlushPriority_Context; + +static ScoreTqReqFlushPriority_Context + ScoreTqReqFlushPriority_Instance; + +static const char * const ScoreTqReqFlushPriority_PreDesc_Queue[] = { + "Empty", + "NonEmpty", + "NA" +}; + +static const char * const * const ScoreTqReqFlushPriority_PreDesc[] = { + ScoreTqReqFlushPriority_PreDesc_Queue, + NULL +}; + +typedef ScoreTqReqFlushPriority_Context Context; + +static const T_scheduler_event *GetUnblock( Context *ctx, size_t *index ) +{ + return TQGetNextUnblock( ctx->tq_ctx, index ); +} + +static const rtems_tcb *GetTCB( Context *ctx, TQWorkerKind worker ) +{ + return ctx->tq_ctx->worker_tcb[ worker ]; +} + +static void Flush( void *arg ) +{ + Context *ctx; + + ctx = arg; + TQSchedulerRecordStart( ctx->tq_ctx ); + TQFlush( ctx->tq_ctx, true ); +} + +static void SchedulerEvent( + void *arg, + const T_scheduler_event *event, + T_scheduler_when when +) +{ + Context *ctx; + + ctx = arg; + + if ( + when == T_SCHEDULER_BEFORE && + event->operation == T_SCHEDULER_BLOCK + ) { + ctx->request.handler = Flush; + ctx->request.arg = ctx; + CallWithinISRSubmit( &ctx->request ); + T_scheduler_set_event_handler( NULL, NULL ); + } +} + +static void ScoreTqReqFlushPriority_Pre_Queue_Prepare( + ScoreTqReqFlushPriority_Context *ctx, + ScoreTqReqFlushPriority_Pre_Queue state +) +{ + switch ( state ) { + case ScoreTqReqFlushPriority_Pre_Queue_Empty: { + /* + * While the thread queue is empty. + */ + ctx->tq_ctx->how_many = 0; + break; + } + + case ScoreTqReqFlushPriority_Pre_Queue_NonEmpty: { + /* + * While the thread queue has at least one enqueued thread. + */ + ctx->tq_ctx->how_many = 3; + break; + } + + case ScoreTqReqFlushPriority_Pre_Queue_NA: + break; + } +} + +static void ScoreTqReqFlushPriority_Post_Operation_Check( + ScoreTqReqFlushPriority_Context *ctx, + ScoreTqReqFlushPriority_Post_Operation state +) +{ + size_t i; + const T_scheduler_event *event; + + i = 0; + + switch ( state ) { + case ScoreTqReqFlushPriority_Post_Operation_Nop: { + /* + * No operation shall be performed. + */ + /* Event receive */ + T_eq_ptr( GetUnblock( ctx, &i )->thread, GetTCB( ctx, TQ_BLOCKER_A ) ); + T_eq_ptr( GetUnblock( ctx, &i ), &T_scheduler_event_null ); + break; + } + + case ScoreTqReqFlushPriority_Post_Operation_TryExtract: { + /* + * The enqueued threads of the thread queue may be extracted in priority + * order for each priority queue associated with a scheduler. The + * priority queues of the thread queue shall be accessed in FIFO order. + */ + event = GetUnblock( ctx, &i ); + T_eq_ptr( event->executing, NULL ); + T_eq_ptr( event->thread, GetTCB( ctx, TQ_BLOCKER_D ) ); + + event = GetUnblock( ctx, &i ); + T_eq_ptr( event->executing, NULL ); + T_eq_ptr( event->thread, GetTCB( ctx, TQ_BLOCKER_C ) ); + + event = GetUnblock( ctx, &i ); + T_eq_ptr( event->executing, GetTCB( ctx, TQ_BLOCKER_B ) ); + T_eq_ptr( event->thread, GetTCB( ctx, TQ_BLOCKER_B ) ); + + T_eq_ptr( GetUnblock( ctx, &i ), &T_scheduler_event_null ); + break; + } + + case ScoreTqReqFlushPriority_Post_Operation_NA: + break; + } +} + +static void ScoreTqReqFlushPriority_Setup( + ScoreTqReqFlushPriority_Context *ctx +) +{ + TQReset( ctx->tq_ctx ); + TQSetPriority( ctx->tq_ctx, TQ_BLOCKER_A, PRIO_ULTRA_HIGH ); + + if ( + ctx->supports_multiple_priority_queues && + rtems_configuration_get_maximum_processors() > 1 + ) { + TQSetScheduler( ctx->tq_ctx, TQ_BLOCKER_B, SCHEDULER_B_ID, PRIO_HIGH ); + } else { + TQSetPriority( ctx->tq_ctx, TQ_BLOCKER_B, PRIO_HIGH ); + } + + TQSetPriority( ctx->tq_ctx, TQ_BLOCKER_C, PRIO_VERY_HIGH ); + TQSetPriority( ctx->tq_ctx, TQ_BLOCKER_D, PRIO_ULTRA_HIGH ); +} + +static void ScoreTqReqFlushPriority_Setup_Wrap( void *arg ) +{ + ScoreTqReqFlushPriority_Context *ctx; + + ctx = arg; + ctx->Map.in_action_loop = false; + ScoreTqReqFlushPriority_Setup( ctx ); +} + +static void ScoreTqReqFlushPriority_Teardown( + ScoreTqReqFlushPriority_Context *ctx +) +{ + TQReset( ctx->tq_ctx ); +} + +static void ScoreTqReqFlushPriority_Teardown_Wrap( void *arg ) +{ + ScoreTqReqFlushPriority_Context *ctx; + + ctx = arg; + ctx->Map.in_action_loop = false; + ScoreTqReqFlushPriority_Teardown( ctx ); +} + +static void ScoreTqReqFlushPriority_Action( + ScoreTqReqFlushPriority_Context *ctx +) +{ + TQSend( ctx->tq_ctx, TQ_BLOCKER_A, TQ_EVENT_ENQUEUE_PREPARE ); + + if ( ctx->tq_ctx->how_many > 0 ) { + TQSend( ctx->tq_ctx, TQ_BLOCKER_C, TQ_EVENT_ENQUEUE ); + TQSend( ctx->tq_ctx, TQ_BLOCKER_D, TQ_EVENT_ENQUEUE ); + T_scheduler_set_event_handler( SchedulerEvent, ctx ); + TQSendAndWaitForExecutionStop( + ctx->tq_ctx, + TQ_BLOCKER_B, + TQ_EVENT_ENQUEUE + ); + } else { + TQSchedulerRecordStart( ctx->tq_ctx ); + TQSend( ctx->tq_ctx, TQ_BLOCKER_A, TQ_EVENT_FLUSH_ALL ); + } + + TQSchedulerRecordStop( ctx->tq_ctx ); + TQSend( ctx->tq_ctx, TQ_BLOCKER_A, TQ_EVENT_ENQUEUE_DONE ); +} + +static const ScoreTqReqFlushPriority_Entry +ScoreTqReqFlushPriority_Entries[] = { + { 0, 0, ScoreTqReqFlushPriority_Post_Operation_Nop }, + { 0, 0, ScoreTqReqFlushPriority_Post_Operation_TryExtract } +}; + +static const uint8_t +ScoreTqReqFlushPriority_Map[] = { + 0, 1 +}; + +static size_t ScoreTqReqFlushPriority_Scope( void *arg, char *buf, size_t n ) +{ + ScoreTqReqFlushPriority_Context *ctx; + + ctx = arg; + + if ( ctx->Map.in_action_loop ) { + return T_get_scope( + ScoreTqReqFlushPriority_PreDesc, + buf, + n, + ctx->Map.pcs + ); + } + + return 0; +} + +static T_fixture ScoreTqReqFlushPriority_Fixture = { + .setup = ScoreTqReqFlushPriority_Setup_Wrap, + .stop = NULL, + .teardown = ScoreTqReqFlushPriority_Teardown_Wrap, + .scope = ScoreTqReqFlushPriority_Scope, + .initial_context = &ScoreTqReqFlushPriority_Instance +}; + +static inline ScoreTqReqFlushPriority_Entry ScoreTqReqFlushPriority_PopEntry( + ScoreTqReqFlushPriority_Context *ctx +) +{ + size_t index; + + index = ctx->Map.index; + ctx->Map.index = index + 1; + return ScoreTqReqFlushPriority_Entries[ + ScoreTqReqFlushPriority_Map[ index ] + ]; +} + +static void ScoreTqReqFlushPriority_TestVariant( + ScoreTqReqFlushPriority_Context *ctx +) +{ + ScoreTqReqFlushPriority_Pre_Queue_Prepare( ctx, ctx->Map.pcs[ 0 ] ); + ScoreTqReqFlushPriority_Action( ctx ); + ScoreTqReqFlushPriority_Post_Operation_Check( + ctx, + ctx->Map.entry.Post_Operation + ); +} + +static T_fixture_node ScoreTqReqFlushPriority_Node; + +static T_remark ScoreTqReqFlushPriority_Remark = { + .next = NULL, + .remark = "ScoreTqReqFlushPriority" +}; + +void ScoreTqReqFlushPriority_Run( + TQContext *tq_ctx, + bool supports_multiple_priority_queues +) +{ + ScoreTqReqFlushPriority_Context *ctx; + + ctx = &ScoreTqReqFlushPriority_Instance; + ctx->tq_ctx = tq_ctx; + ctx->supports_multiple_priority_queues = supports_multiple_priority_queues; + + ctx = T_push_fixture( + &ScoreTqReqFlushPriority_Node, + &ScoreTqReqFlushPriority_Fixture + ); + ctx->Map.in_action_loop = true; + ctx->Map.index = 0; + + for ( + ctx->Map.pcs[ 0 ] = ScoreTqReqFlushPriority_Pre_Queue_Empty; + ctx->Map.pcs[ 0 ] < ScoreTqReqFlushPriority_Pre_Queue_NA; + ++ctx->Map.pcs[ 0 ] + ) { + ctx->Map.entry = ScoreTqReqFlushPriority_PopEntry( ctx ); + ScoreTqReqFlushPriority_TestVariant( ctx ); + } + + T_add_remark( &ScoreTqReqFlushPriority_Remark ); + T_pop_fixture(); +} + +/** @} */ |