diff options
author | Sebastian Huber <sebastian.huber@embedded-brains.de> | 2021-08-31 11:09:43 +0200 |
---|---|---|
committer | Sebastian Huber <sebastian.huber@embedded-brains.de> | 2021-09-24 13:27:22 +0200 |
commit | 5d1211d18182948959d7be4a5618b310be087456 (patch) | |
tree | de68812112995cd3e64aeb23e29672fb9c00bf7c | |
parent | 4022a76e95ad3519b38957cfc3ad70974607a644 (diff) |
validation: Thread queue flush runner
-rw-r--r-- | spec/build/testsuites/validation/validation-0.yml | 3 | ||||
-rw-r--r-- | testsuites/validation/tr-tq-flush-fifo.c | 381 | ||||
-rw-r--r-- | testsuites/validation/tr-tq-flush-fifo.h | 91 | ||||
-rw-r--r-- | testsuites/validation/tr-tq-flush-priority-inherit.c | 571 | ||||
-rw-r--r-- | testsuites/validation/tr-tq-flush-priority-inherit.h | 103 | ||||
-rw-r--r-- | testsuites/validation/tr-tq-flush-priority.c | 419 | ||||
-rw-r--r-- | testsuites/validation/tr-tq-flush-priority.h | 97 |
7 files changed, 1665 insertions, 0 deletions
diff --git a/spec/build/testsuites/validation/validation-0.yml b/spec/build/testsuites/validation/validation-0.yml index 3a5d2d6739..f8de87dd64 100644 --- a/spec/build/testsuites/validation/validation-0.yml +++ b/spec/build/testsuites/validation/validation-0.yml @@ -116,6 +116,9 @@ source: - testsuites/validation/tr-tq-enqueue-fifo.c - testsuites/validation/tr-tq-enqueue-mrsp.c - testsuites/validation/tr-tq-enqueue-priority.c +- testsuites/validation/tr-tq-flush-fifo.c +- testsuites/validation/tr-tq-flush-priority.c +- testsuites/validation/tr-tq-flush-priority-inherit.c - testsuites/validation/ts-validation-0.c stlib: [] target: testsuites/validation/ts-validation-0.exe diff --git a/testsuites/validation/tr-tq-flush-fifo.c b/testsuites/validation/tr-tq-flush-fifo.c new file mode 100644 index 0000000000..70287000a8 --- /dev/null +++ b/testsuites/validation/tr-tq-flush-fifo.c @@ -0,0 +1,381 @@ +/* SPDX-License-Identifier: BSD-2-Clause */ + +/** + * @file + * + * @ingroup RTEMSTestCaseScoreTqReqFlushFifo + */ + +/* + * Copyright (C) 2021 embedded brains GmbH (http://www.embedded-brains.de) + * + * Redistribution and use in source and binary forms, with or without + * modification, are permitted provided that the following conditions + * are met: + * 1. Redistributions of source code must retain the above copyright + * notice, this list of conditions and the following disclaimer. + * 2. Redistributions in binary form must reproduce the above copyright + * notice, this list of conditions and the following disclaimer in the + * documentation and/or other materials provided with the distribution. + * + * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" + * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE + * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE + * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE + * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR + * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF + * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS + * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN + * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) + * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE + * POSSIBILITY OF SUCH DAMAGE. + */ + +/* + * This file is part of the RTEMS quality process and was automatically + * generated. If you find something that needs to be fixed or + * worded better please post a report or patch to an RTEMS mailing list + * or raise a bug report: + * + * https://www.rtems.org/bugs.html + * + * For information on updating and regenerating please refer to the How-To + * section in the Software Requirements Engineering chapter of the + * RTEMS Software Engineering manual. The manual is provided as a part of + * a release. For development sources please refer to the online + * documentation at: + * + * https://docs.rtems.org + */ + +#ifdef HAVE_CONFIG_H +#include "config.h" +#endif + +#include "tr-tq-flush-fifo.h" +#include "tx-support.h" + +#include <rtems/test.h> + +/** + * @defgroup RTEMSTestCaseScoreTqReqFlushFifo spec:/score/tq/req/flush-fifo + * + * @ingroup RTEMSTestSuiteTestsuitesValidation0 + * + * @{ + */ + +typedef struct { + uint8_t Skip : 1; + uint8_t Pre_Queue_NA : 1; + uint8_t Post_Operation : 2; +} ScoreTqReqFlushFifo_Entry; + +/** + * @brief Test context for spec:/score/tq/req/flush-fifo test case. + */ +typedef struct { + /** + * @brief This member contains the call within ISR request. + */ + CallWithinISRRequest request;; + + /** + * @brief This member contains a copy of the corresponding + * ScoreTqReqFlushFifo_Run() parameter. + */ + TQContext *tq_ctx; + + struct { + /** + * @brief This member defines the pre-condition states for the next action. + */ + size_t pcs[ 1 ]; + + /** + * @brief If this member is true, then the test action loop is executed. + */ + bool in_action_loop; + + /** + * @brief This member contains the next transition map index. + */ + size_t index; + + /** + * @brief This member contains the current transition map entry. + */ + ScoreTqReqFlushFifo_Entry entry; + + /** + * @brief If this member is true, then the current transition variant + * should be skipped. + */ + bool skip; + } Map; +} ScoreTqReqFlushFifo_Context; + +static ScoreTqReqFlushFifo_Context + ScoreTqReqFlushFifo_Instance; + +static const char * const ScoreTqReqFlushFifo_PreDesc_Queue[] = { + "Empty", + "NonEmpty", + "NA" +}; + +static const char * const * const ScoreTqReqFlushFifo_PreDesc[] = { + ScoreTqReqFlushFifo_PreDesc_Queue, + NULL +}; + +typedef ScoreTqReqFlushFifo_Context Context; + +static const T_scheduler_event *GetUnblock( Context *ctx, size_t *index ) +{ + return TQGetNextUnblock( ctx->tq_ctx, index ); +} + +static const rtems_tcb *GetTCB( Context *ctx, TQWorkerKind worker ) +{ + return ctx->tq_ctx->worker_tcb[ worker ]; +} + +static void Flush( void *arg ) +{ + Context *ctx; + + ctx = arg; + TQSchedulerRecordStart( ctx->tq_ctx ); + TQFlush( ctx->tq_ctx ); +} + +static void SchedulerEvent( + void *arg, + const T_scheduler_event *event, + T_scheduler_when when +) +{ + Context *ctx; + + ctx = arg; + + if ( + when == T_SCHEDULER_BEFORE && + event->operation == T_SCHEDULER_BLOCK + ) { + ctx->request.handler = Flush; + ctx->request.arg = ctx; + CallWithinISRSubmit( &ctx->request ); + T_scheduler_set_event_handler( NULL, NULL ); + } +} + +static void ScoreTqReqFlushFifo_Pre_Queue_Prepare( + ScoreTqReqFlushFifo_Context *ctx, + ScoreTqReqFlushFifo_Pre_Queue state +) +{ + switch ( state ) { + case ScoreTqReqFlushFifo_Pre_Queue_Empty: { + /* + * While the thread queue is empty. + */ + ctx->tq_ctx->how_many = 0; + break; + } + + case ScoreTqReqFlushFifo_Pre_Queue_NonEmpty: { + /* + * While the thread queue has at least one enqueued thread. + */ + ctx->tq_ctx->how_many = 3; + break; + } + + case ScoreTqReqFlushFifo_Pre_Queue_NA: + break; + } +} + +static void ScoreTqReqFlushFifo_Post_Operation_Check( + ScoreTqReqFlushFifo_Context *ctx, + ScoreTqReqFlushFifo_Post_Operation state +) +{ + size_t i; + const T_scheduler_event *event; + + i = 0; + + switch ( state ) { + case ScoreTqReqFlushFifo_Post_Operation_Nop: { + /* + * No operation shall be performed. + */ + /* Event receive */ + T_eq_ptr( GetUnblock( ctx, &i )->thread, GetTCB( ctx, TQ_BLOCKER_A ) ); + T_eq_ptr( GetUnblock( ctx, &i ), &T_scheduler_event_null ); + break; + } + + case ScoreTqReqFlushFifo_Post_Operation_TryExtract: { + /* + * The enqueued threads of the thread queue may be extracted in FIFO + * order. + */ + event = GetUnblock( ctx, &i ); + T_eq_ptr( event->executing, NULL ); + T_eq_ptr( event->thread, GetTCB( ctx, TQ_BLOCKER_B ) ); + + event = GetUnblock( ctx, &i ); + T_eq_ptr( event->executing, NULL ); + T_eq_ptr( event->thread, GetTCB( ctx, TQ_BLOCKER_C ) ); + + event = GetUnblock( ctx, &i ); + T_eq_ptr( event->executing, GetTCB( ctx, TQ_BLOCKER_D ) ); + T_eq_ptr( event->thread, GetTCB( ctx, TQ_BLOCKER_D ) ); + + T_eq_ptr( GetUnblock( ctx, &i ), &T_scheduler_event_null ); + break; + } + + case ScoreTqReqFlushFifo_Post_Operation_NA: + break; + } +} + +static void ScoreTqReqFlushFifo_Setup( ScoreTqReqFlushFifo_Context *ctx ) +{ + TQReset( ctx->tq_ctx ); + TQSetPriority( ctx->tq_ctx, TQ_BLOCKER_A, PRIO_ULTRA_HIGH ); + TQSetPriority( ctx->tq_ctx, TQ_BLOCKER_B, PRIO_VERY_HIGH ); + TQSetPriority( ctx->tq_ctx, TQ_BLOCKER_C, PRIO_HIGH ); + TQSetPriority( ctx->tq_ctx, TQ_BLOCKER_D, PRIO_HIGH ); +} + +static void ScoreTqReqFlushFifo_Setup_Wrap( void *arg ) +{ + ScoreTqReqFlushFifo_Context *ctx; + + ctx = arg; + ctx->Map.in_action_loop = false; + ScoreTqReqFlushFifo_Setup( ctx ); +} + +static void ScoreTqReqFlushFifo_Teardown( ScoreTqReqFlushFifo_Context *ctx ) +{ + TQReset( ctx->tq_ctx ); +} + +static void ScoreTqReqFlushFifo_Teardown_Wrap( void *arg ) +{ + ScoreTqReqFlushFifo_Context *ctx; + + ctx = arg; + ctx->Map.in_action_loop = false; + ScoreTqReqFlushFifo_Teardown( ctx ); +} + +static void ScoreTqReqFlushFifo_Action( ScoreTqReqFlushFifo_Context *ctx ) +{ + TQSend( ctx->tq_ctx, TQ_BLOCKER_A, TQ_EVENT_ENQUEUE_PREPARE ); + + if ( ctx->tq_ctx->how_many > 0 ) { + TQSend( ctx->tq_ctx, TQ_BLOCKER_B, TQ_EVENT_ENQUEUE ); + TQSend( ctx->tq_ctx, TQ_BLOCKER_C, TQ_EVENT_ENQUEUE ); + T_scheduler_set_event_handler( SchedulerEvent, ctx ); + TQSend( ctx->tq_ctx, TQ_BLOCKER_D, TQ_EVENT_ENQUEUE ); + } else { + TQSchedulerRecordStart( ctx->tq_ctx ); + TQSend( ctx->tq_ctx, TQ_BLOCKER_A, TQ_EVENT_FLUSH ); + } + + TQSchedulerRecordStop( ctx->tq_ctx ); + TQSend( ctx->tq_ctx, TQ_BLOCKER_A, TQ_EVENT_ENQUEUE_DONE ); +} + +static const ScoreTqReqFlushFifo_Entry +ScoreTqReqFlushFifo_Entries[] = { + { 0, 0, ScoreTqReqFlushFifo_Post_Operation_Nop }, + { 0, 0, ScoreTqReqFlushFifo_Post_Operation_TryExtract } +}; + +static const uint8_t +ScoreTqReqFlushFifo_Map[] = { + 0, 1 +}; + +static size_t ScoreTqReqFlushFifo_Scope( void *arg, char *buf, size_t n ) +{ + ScoreTqReqFlushFifo_Context *ctx; + + ctx = arg; + + if ( ctx->Map.in_action_loop ) { + return T_get_scope( ScoreTqReqFlushFifo_PreDesc, buf, n, ctx->Map.pcs ); + } + + return 0; +} + +static T_fixture ScoreTqReqFlushFifo_Fixture = { + .setup = ScoreTqReqFlushFifo_Setup_Wrap, + .stop = NULL, + .teardown = ScoreTqReqFlushFifo_Teardown_Wrap, + .scope = ScoreTqReqFlushFifo_Scope, + .initial_context = &ScoreTqReqFlushFifo_Instance +}; + +static inline ScoreTqReqFlushFifo_Entry ScoreTqReqFlushFifo_PopEntry( + ScoreTqReqFlushFifo_Context *ctx +) +{ + size_t index; + + index = ctx->Map.index; + ctx->Map.index = index + 1; + return ScoreTqReqFlushFifo_Entries[ + ScoreTqReqFlushFifo_Map[ index ] + ]; +} + +static void ScoreTqReqFlushFifo_TestVariant( ScoreTqReqFlushFifo_Context *ctx ) +{ + ScoreTqReqFlushFifo_Pre_Queue_Prepare( ctx, ctx->Map.pcs[ 0 ] ); + ScoreTqReqFlushFifo_Action( ctx ); + ScoreTqReqFlushFifo_Post_Operation_Check( + ctx, + ctx->Map.entry.Post_Operation + ); +} + +static T_fixture_node ScoreTqReqFlushFifo_Node; + +void ScoreTqReqFlushFifo_Run( TQContext *tq_ctx ) +{ + ScoreTqReqFlushFifo_Context *ctx; + + ctx = &ScoreTqReqFlushFifo_Instance; + ctx->tq_ctx = tq_ctx; + + ctx = T_push_fixture( + &ScoreTqReqFlushFifo_Node, + &ScoreTqReqFlushFifo_Fixture + ); + ctx->Map.in_action_loop = true; + ctx->Map.index = 0; + + for ( + ctx->Map.pcs[ 0 ] = ScoreTqReqFlushFifo_Pre_Queue_Empty; + ctx->Map.pcs[ 0 ] < ScoreTqReqFlushFifo_Pre_Queue_NA; + ++ctx->Map.pcs[ 0 ] + ) { + ctx->Map.entry = ScoreTqReqFlushFifo_PopEntry( ctx ); + ScoreTqReqFlushFifo_TestVariant( ctx ); + } + + T_pop_fixture(); +} + +/** @} */ diff --git a/testsuites/validation/tr-tq-flush-fifo.h b/testsuites/validation/tr-tq-flush-fifo.h new file mode 100644 index 0000000000..dca1444891 --- /dev/null +++ b/testsuites/validation/tr-tq-flush-fifo.h @@ -0,0 +1,91 @@ +/* SPDX-License-Identifier: BSD-2-Clause */ + +/** + * @file + * + * @ingroup RTEMSTestCaseScoreTqReqFlushFifo + */ + +/* + * Copyright (C) 2021 embedded brains GmbH (http://www.embedded-brains.de) + * + * Redistribution and use in source and binary forms, with or without + * modification, are permitted provided that the following conditions + * are met: + * 1. Redistributions of source code must retain the above copyright + * notice, this list of conditions and the following disclaimer. + * 2. Redistributions in binary form must reproduce the above copyright + * notice, this list of conditions and the following disclaimer in the + * documentation and/or other materials provided with the distribution. + * + * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" + * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE + * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE + * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE + * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR + * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF + * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS + * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN + * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) + * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE + * POSSIBILITY OF SUCH DAMAGE. + */ + +/* + * This file is part of the RTEMS quality process and was automatically + * generated. If you find something that needs to be fixed or + * worded better please post a report or patch to an RTEMS mailing list + * or raise a bug report: + * + * https://www.rtems.org/bugs.html + * + * For information on updating and regenerating please refer to the How-To + * section in the Software Requirements Engineering chapter of the + * RTEMS Software Engineering manual. The manual is provided as a part of + * a release. For development sources please refer to the online + * documentation at: + * + * https://docs.rtems.org + */ + +#ifndef _TR_TQ_FLUSH_FIFO_H +#define _TR_TQ_FLUSH_FIFO_H + +#include "tx-thread-queue.h" + +#ifdef __cplusplus +extern "C" { +#endif + +/** + * @addtogroup RTEMSTestCaseScoreTqReqFlushFifo + * + * @{ + */ + +typedef enum { + ScoreTqReqFlushFifo_Pre_Queue_Empty, + ScoreTqReqFlushFifo_Pre_Queue_NonEmpty, + ScoreTqReqFlushFifo_Pre_Queue_NA +} ScoreTqReqFlushFifo_Pre_Queue; + +typedef enum { + ScoreTqReqFlushFifo_Post_Operation_Nop, + ScoreTqReqFlushFifo_Post_Operation_TryExtract, + ScoreTqReqFlushFifo_Post_Operation_NA +} ScoreTqReqFlushFifo_Post_Operation; + +/** + * @brief Runs the parameterized test case. + * + * @param[in,out] tq_ctx is the thread queue test context. + */ +void ScoreTqReqFlushFifo_Run( TQContext *tq_ctx ); + +/** @} */ + +#ifdef __cplusplus +} +#endif + +#endif /* _TR_TQ_FLUSH_FIFO_H */ diff --git a/testsuites/validation/tr-tq-flush-priority-inherit.c b/testsuites/validation/tr-tq-flush-priority-inherit.c new file mode 100644 index 0000000000..9031fba65f --- /dev/null +++ b/testsuites/validation/tr-tq-flush-priority-inherit.c @@ -0,0 +1,571 @@ +/* SPDX-License-Identifier: BSD-2-Clause */ + +/** + * @file + * + * @ingroup RTEMSTestCaseScoreTqReqFlushPriorityInherit + */ + +/* + * Copyright (C) 2021 embedded brains GmbH (http://www.embedded-brains.de) + * + * Redistribution and use in source and binary forms, with or without + * modification, are permitted provided that the following conditions + * are met: + * 1. Redistributions of source code must retain the above copyright + * notice, this list of conditions and the following disclaimer. + * 2. Redistributions in binary form must reproduce the above copyright + * notice, this list of conditions and the following disclaimer in the + * documentation and/or other materials provided with the distribution. + * + * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" + * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE + * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE + * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE + * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR + * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF + * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS + * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN + * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) + * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE + * POSSIBILITY OF SUCH DAMAGE. + */ + +/* + * This file is part of the RTEMS quality process and was automatically + * generated. If you find something that needs to be fixed or + * worded better please post a report or patch to an RTEMS mailing list + * or raise a bug report: + * + * https://www.rtems.org/bugs.html + * + * For information on updating and regenerating please refer to the How-To + * section in the Software Requirements Engineering chapter of the + * RTEMS Software Engineering manual. The manual is provided as a part of + * a release. For development sources please refer to the online + * documentation at: + * + * https://docs.rtems.org + */ + +#ifdef HAVE_CONFIG_H +#include "config.h" +#endif + +#include "tr-tq-flush-priority-inherit.h" +#include "tx-support.h" + +#include <rtems/test.h> + +/** + * @defgroup RTEMSTestCaseScoreTqReqFlushPriorityInherit \ + * spec:/score/tq/req/flush-priority-inherit + * + * @ingroup RTEMSTestSuiteTestsuitesValidation0 + * + * @{ + */ + +typedef struct { + uint8_t Skip : 1; + uint8_t Pre_Priority_NA : 1; + uint8_t Pre_Queue_NA : 1; + uint8_t Post_Extract : 2; + uint8_t Post_PriorityUpdate : 2; +} ScoreTqReqFlushPriorityInherit_Entry; + +/** + * @brief Test context for spec:/score/tq/req/flush-priority-inherit test case. + */ +typedef struct { + /** + * @brief This member contains the call within ISR request. + */ + CallWithinISRRequest request;; + + /** + * @brief If this member is true, then a minimum priority of the owner of the + * thread queue shall be inherited from a thread enqueued on the thread + * queue. + */ + bool minimum; + + /** + * @brief This member contains a copy of the corresponding + * ScoreTqReqFlushPriorityInherit_Run() parameter. + */ + TQContext *tq_ctx; + + struct { + /** + * @brief This member defines the pre-condition indices for the next + * action. + */ + size_t pci[ 2 ]; + + /** + * @brief This member defines the pre-condition states for the next action. + */ + size_t pcs[ 2 ]; + + /** + * @brief If this member is true, then the test action loop is executed. + */ + bool in_action_loop; + + /** + * @brief This member contains the next transition map index. + */ + size_t index; + + /** + * @brief This member contains the current transition map entry. + */ + ScoreTqReqFlushPriorityInherit_Entry entry; + + /** + * @brief If this member is true, then the current transition variant + * should be skipped. + */ + bool skip; + } Map; +} ScoreTqReqFlushPriorityInherit_Context; + +static ScoreTqReqFlushPriorityInherit_Context + ScoreTqReqFlushPriorityInherit_Instance; + +static const char * const ScoreTqReqFlushPriorityInherit_PreDesc_Priority[] = { + "Minimum", + "NotMinimum", + "NA" +}; + +static const char * const ScoreTqReqFlushPriorityInherit_PreDesc_Queue[] = { + "Empty", + "NonEmpty", + "NA" +}; + +static const char * const * const ScoreTqReqFlushPriorityInherit_PreDesc[] = { + ScoreTqReqFlushPriorityInherit_PreDesc_Priority, + ScoreTqReqFlushPriorityInherit_PreDesc_Queue, + NULL +}; + +typedef ScoreTqReqFlushPriorityInherit_Context Context; + +static const T_scheduler_event *GetUnblock( Context *ctx, size_t *index ) +{ + return TQGetNextUnblock( ctx->tq_ctx, index ); +} + +static const T_scheduler_event *GetPriorityUpdate( Context *ctx, size_t *index ) +{ + return T_scheduler_next( + &ctx->tq_ctx->scheduler_log.header, + T_SCHEDULER_UPDATE_PRIORITY, + index + ); +} + +static const rtems_tcb *GetTCB( Context *ctx, TQWorkerKind worker ) +{ + return ctx->tq_ctx->worker_tcb[ worker ]; +} + +static void Flush( void *arg ) +{ + Context *ctx; + + ctx = arg; + TQSchedulerRecordStart( ctx->tq_ctx ); + TQFlush( ctx->tq_ctx ); +} + +static void SchedulerEvent( + void *arg, + const T_scheduler_event *event, + T_scheduler_when when +) +{ + Context *ctx; + + ctx = arg; + + if ( + when == T_SCHEDULER_BEFORE && + event->operation == T_SCHEDULER_BLOCK + ) { + ctx->request.handler = Flush; + ctx->request.arg = ctx; + CallWithinISRSubmit( &ctx->request ); + T_scheduler_set_event_handler( NULL, NULL ); + } +} + +static void ScoreTqReqFlushPriorityInherit_Pre_Priority_Prepare( + ScoreTqReqFlushPriorityInherit_Context *ctx, + ScoreTqReqFlushPriorityInherit_Pre_Priority state +) +{ + switch ( state ) { + case ScoreTqReqFlushPriorityInherit_Pre_Priority_Minimum: { + /* + * While a minimum priority of the owner of the thread queue is inherited + * from a thread enqueued on the thread queue. + */ + ctx->minimum = true; + break; + } + + case ScoreTqReqFlushPriorityInherit_Pre_Priority_NotMinimum: { + /* + * While no minimum priority of the owner of the thread queue is + * inherited from a thread enqueued on the thread queue. + */ + ctx->minimum = false; + break; + } + + case ScoreTqReqFlushPriorityInherit_Pre_Priority_NA: + break; + } +} + +static void ScoreTqReqFlushPriorityInherit_Pre_Queue_Prepare( + ScoreTqReqFlushPriorityInherit_Context *ctx, + ScoreTqReqFlushPriorityInherit_Pre_Queue state +) +{ + switch ( state ) { + case ScoreTqReqFlushPriorityInherit_Pre_Queue_Empty: { + /* + * While the thread queue is empty. + */ + ctx->tq_ctx->how_many = 0; + break; + } + + case ScoreTqReqFlushPriorityInherit_Pre_Queue_NonEmpty: { + /* + * While the thread queue has at least one enqueued thread. + */ + ctx->tq_ctx->how_many = 3; + break; + } + + case ScoreTqReqFlushPriorityInherit_Pre_Queue_NA: + break; + } +} + +static void ScoreTqReqFlushPriorityInherit_Post_Extract_Check( + ScoreTqReqFlushPriorityInherit_Context *ctx, + ScoreTqReqFlushPriorityInherit_Post_Extract state +) +{ + size_t i; + const T_scheduler_event *event; + + i = 0; + + switch ( state ) { + case ScoreTqReqFlushPriorityInherit_Post_Extract_Nop: { + /* + * No operation shall be performed. + */ + /* Event receive */ + T_eq_ptr( GetUnblock( ctx, &i )->thread, GetTCB( ctx, TQ_BLOCKER_A ) ); + T_eq_ptr( GetPriorityUpdate( ctx, &i ), &T_scheduler_event_null ); + break; + } + + case ScoreTqReqFlushPriorityInherit_Post_Extract_All: { + /* + * The enqueued threads of the thread queue shall be extracted in + * priority order for each priority queue associated with a scheduler. + * The priority queues of the thread queue shall be accessed in FIFO + * order. + */ + event = GetUnblock( ctx, &i ); + T_eq_ptr( event->executing, NULL ); + T_eq_ptr( event->thread, GetTCB( ctx, TQ_BLOCKER_D ) ); + + event = GetUnblock( ctx, &i ); + T_eq_ptr( event->executing, NULL ); + T_eq_ptr( event->thread, GetTCB( ctx, TQ_BLOCKER_C ) ); + + if ( ctx->minimum ) { + /* + * This priority update is carried out by + * _Thread_queue_Flush_critical(). + */ + event = GetPriorityUpdate( ctx, &i ); + T_eq_ptr( event->executing, NULL ); + T_eq_ptr( event->thread, GetTCB( ctx, TQ_BLOCKER_A ) ); + } + + event = GetUnblock( ctx, &i ); + T_eq_ptr( event->executing, GetTCB( ctx, TQ_BLOCKER_B ) ); + T_eq_ptr( event->thread, GetTCB( ctx, TQ_BLOCKER_B ) ); + + if ( ctx->minimum && rtems_configuration_get_maximum_processors() > 1 ) { + /* + * This superfluous priority update is carried out by + * _Thread_queue_Enqueue() since TQ_BLOCKER_B would have inherited its + * priority for scheduler B to TQ_BLOCKER_A if it were not flushed from + * the thread queue. + */ + event = GetPriorityUpdate( ctx, &i ); + T_eq_ptr( event->executing, GetTCB( ctx, TQ_BLOCKER_B ) ); + T_eq_ptr( event->thread, GetTCB( ctx, TQ_BLOCKER_A ) ); + } + + event = GetPriorityUpdate( ctx, &i ); + T_eq_ptr( event, &T_scheduler_event_null ); + + T_eq_u32( + GetPriority( ctx->tq_ctx->worker_id[ TQ_BLOCKER_A ] ), + PRIO_HIGH + ); + break; + } + + case ScoreTqReqFlushPriorityInherit_Post_Extract_NA: + break; + } +} + +static void ScoreTqReqFlushPriorityInherit_Post_PriorityUpdate_Check( + ScoreTqReqFlushPriorityInherit_Context *ctx, + ScoreTqReqFlushPriorityInherit_Post_PriorityUpdate state +) +{ + switch ( state ) { + case ScoreTqReqFlushPriorityInherit_Post_PriorityUpdate_No: { + /* + * The current priority of the owner of the thread queue shall not be + * updated by the thread queue flush operation. + */ + /* Checked by ``Extract`` post-condition state ``Nop`` */ + break; + } + + case ScoreTqReqFlushPriorityInherit_Post_PriorityUpdate_Yes: { + /* + * The current priority of the owner of the thread queue shall be updated + * by the thread queue flush operation to reflect the loss of inherited + * priorities of the flushed threads. + */ + /* Checked by ``Extract`` post-condition state ``All`` */ + break; + } + + case ScoreTqReqFlushPriorityInherit_Post_PriorityUpdate_NA: + break; + } +} + +static void ScoreTqReqFlushPriorityInherit_Setup( + ScoreTqReqFlushPriorityInherit_Context *ctx +) +{ + TQReset( ctx->tq_ctx ); + TQSetPriority( ctx->tq_ctx, TQ_BLOCKER_A, PRIO_HIGH ); +} + +static void ScoreTqReqFlushPriorityInherit_Setup_Wrap( void *arg ) +{ + ScoreTqReqFlushPriorityInherit_Context *ctx; + + ctx = arg; + ctx->Map.in_action_loop = false; + ScoreTqReqFlushPriorityInherit_Setup( ctx ); +} + +static void ScoreTqReqFlushPriorityInherit_Teardown( + ScoreTqReqFlushPriorityInherit_Context *ctx +) +{ + TQReset( ctx->tq_ctx ); +} + +static void ScoreTqReqFlushPriorityInherit_Teardown_Wrap( void *arg ) +{ + ScoreTqReqFlushPriorityInherit_Context *ctx; + + ctx = arg; + ctx->Map.in_action_loop = false; + ScoreTqReqFlushPriorityInherit_Teardown( ctx ); +} + +static void ScoreTqReqFlushPriorityInherit_Action( + ScoreTqReqFlushPriorityInherit_Context *ctx +) +{ + TQSend( ctx->tq_ctx, TQ_BLOCKER_A, TQ_EVENT_ENQUEUE_PREPARE ); + + if ( ctx->tq_ctx->how_many > 0 ) { + if ( ctx->minimum ) { + if ( rtems_configuration_get_maximum_processors() > 1 ) { + TQSetScheduler( ctx->tq_ctx, TQ_BLOCKER_B, SCHEDULER_B_ID, PRIO_HIGH ); + } else { + TQSetScheduler( ctx->tq_ctx, TQ_BLOCKER_B, SCHEDULER_A_ID, PRIO_HIGH ); + } + + TQSetPriority( ctx->tq_ctx, TQ_BLOCKER_C, PRIO_VERY_HIGH ); + TQSetPriority( ctx->tq_ctx, TQ_BLOCKER_D, PRIO_ULTRA_HIGH ); + TQSend( ctx->tq_ctx, TQ_BLOCKER_C, TQ_EVENT_ENQUEUE ); + TQSend( ctx->tq_ctx, TQ_BLOCKER_D, TQ_EVENT_ENQUEUE ); + } else { + TQSetScheduler( ctx->tq_ctx, TQ_BLOCKER_B, SCHEDULER_A_ID, PRIO_HIGH ); + TQSetPriority( ctx->tq_ctx, TQ_BLOCKER_C, PRIO_HIGH ); + TQSetPriority( ctx->tq_ctx, TQ_BLOCKER_D, PRIO_HIGH ); + + TQSend( ctx->tq_ctx, TQ_BLOCKER_D, TQ_EVENT_ENQUEUE ); + TQSend( ctx->tq_ctx, TQ_BLOCKER_C, TQ_EVENT_ENQUEUE ); + } + + T_scheduler_set_event_handler( SchedulerEvent, ctx ); + TQSendAndWaitForExecutionStop( + ctx->tq_ctx, + TQ_BLOCKER_B, + TQ_EVENT_ENQUEUE + ); + } else { + TQSchedulerRecordStart( ctx->tq_ctx ); + TQSend( ctx->tq_ctx, TQ_BLOCKER_A, TQ_EVENT_FLUSH ); + } + + TQSchedulerRecordStop( ctx->tq_ctx ); + TQSend( ctx->tq_ctx, TQ_BLOCKER_A, TQ_EVENT_ENQUEUE_DONE ); +} + +static const ScoreTqReqFlushPriorityInherit_Entry +ScoreTqReqFlushPriorityInherit_Entries[] = { + { 0, 1, 0, ScoreTqReqFlushPriorityInherit_Post_Extract_Nop, + ScoreTqReqFlushPriorityInherit_Post_PriorityUpdate_No }, + { 0, 0, 0, ScoreTqReqFlushPriorityInherit_Post_Extract_All, + ScoreTqReqFlushPriorityInherit_Post_PriorityUpdate_Yes }, + { 0, 0, 0, ScoreTqReqFlushPriorityInherit_Post_Extract_All, + ScoreTqReqFlushPriorityInherit_Post_PriorityUpdate_No } +}; + +static const uint8_t +ScoreTqReqFlushPriorityInherit_Map[] = { + 0, 1, 0, 2 +}; + +static size_t ScoreTqReqFlushPriorityInherit_Scope( + void *arg, + char *buf, + size_t n +) +{ + ScoreTqReqFlushPriorityInherit_Context *ctx; + + ctx = arg; + + if ( ctx->Map.in_action_loop ) { + return T_get_scope( + ScoreTqReqFlushPriorityInherit_PreDesc, + buf, + n, + ctx->Map.pcs + ); + } + + return 0; +} + +static T_fixture ScoreTqReqFlushPriorityInherit_Fixture = { + .setup = ScoreTqReqFlushPriorityInherit_Setup_Wrap, + .stop = NULL, + .teardown = ScoreTqReqFlushPriorityInherit_Teardown_Wrap, + .scope = ScoreTqReqFlushPriorityInherit_Scope, + .initial_context = &ScoreTqReqFlushPriorityInherit_Instance +}; + +static inline ScoreTqReqFlushPriorityInherit_Entry +ScoreTqReqFlushPriorityInherit_PopEntry( + ScoreTqReqFlushPriorityInherit_Context *ctx +) +{ + size_t index; + + index = ctx->Map.index; + ctx->Map.index = index + 1; + return ScoreTqReqFlushPriorityInherit_Entries[ + ScoreTqReqFlushPriorityInherit_Map[ index ] + ]; +} + +static void ScoreTqReqFlushPriorityInherit_SetPreConditionStates( + ScoreTqReqFlushPriorityInherit_Context *ctx +) +{ + if ( ctx->Map.entry.Pre_Priority_NA ) { + ctx->Map.pcs[ 0 ] = ScoreTqReqFlushPriorityInherit_Pre_Priority_NA; + } else { + ctx->Map.pcs[ 0 ] = ctx->Map.pci[ 0 ]; + } + + ctx->Map.pcs[ 1 ] = ctx->Map.pci[ 1 ]; +} + +static void ScoreTqReqFlushPriorityInherit_TestVariant( + ScoreTqReqFlushPriorityInherit_Context *ctx +) +{ + ScoreTqReqFlushPriorityInherit_Pre_Priority_Prepare( + ctx, + ctx->Map.pcs[ 0 ] + ); + ScoreTqReqFlushPriorityInherit_Pre_Queue_Prepare( ctx, ctx->Map.pcs[ 1 ] ); + ScoreTqReqFlushPriorityInherit_Action( ctx ); + ScoreTqReqFlushPriorityInherit_Post_Extract_Check( + ctx, + ctx->Map.entry.Post_Extract + ); + ScoreTqReqFlushPriorityInherit_Post_PriorityUpdate_Check( + ctx, + ctx->Map.entry.Post_PriorityUpdate + ); +} + +static T_fixture_node ScoreTqReqFlushPriorityInherit_Node; + +void ScoreTqReqFlushPriorityInherit_Run( TQContext *tq_ctx ) +{ + ScoreTqReqFlushPriorityInherit_Context *ctx; + + ctx = &ScoreTqReqFlushPriorityInherit_Instance; + ctx->tq_ctx = tq_ctx; + + ctx = T_push_fixture( + &ScoreTqReqFlushPriorityInherit_Node, + &ScoreTqReqFlushPriorityInherit_Fixture + ); + ctx->Map.in_action_loop = true; + ctx->Map.index = 0; + + for ( + ctx->Map.pci[ 0 ] = ScoreTqReqFlushPriorityInherit_Pre_Priority_Minimum; + ctx->Map.pci[ 0 ] < ScoreTqReqFlushPriorityInherit_Pre_Priority_NA; + ++ctx->Map.pci[ 0 ] + ) { + for ( + ctx->Map.pci[ 1 ] = ScoreTqReqFlushPriorityInherit_Pre_Queue_Empty; + ctx->Map.pci[ 1 ] < ScoreTqReqFlushPriorityInherit_Pre_Queue_NA; + ++ctx->Map.pci[ 1 ] + ) { + ctx->Map.entry = ScoreTqReqFlushPriorityInherit_PopEntry( ctx ); + ScoreTqReqFlushPriorityInherit_SetPreConditionStates( ctx ); + ScoreTqReqFlushPriorityInherit_TestVariant( ctx ); + } + } + + T_pop_fixture(); +} + +/** @} */ diff --git a/testsuites/validation/tr-tq-flush-priority-inherit.h b/testsuites/validation/tr-tq-flush-priority-inherit.h new file mode 100644 index 0000000000..1b12f1afb0 --- /dev/null +++ b/testsuites/validation/tr-tq-flush-priority-inherit.h @@ -0,0 +1,103 @@ +/* SPDX-License-Identifier: BSD-2-Clause */ + +/** + * @file + * + * @ingroup RTEMSTestCaseScoreTqReqFlushPriorityInherit + */ + +/* + * Copyright (C) 2021 embedded brains GmbH (http://www.embedded-brains.de) + * + * Redistribution and use in source and binary forms, with or without + * modification, are permitted provided that the following conditions + * are met: + * 1. Redistributions of source code must retain the above copyright + * notice, this list of conditions and the following disclaimer. + * 2. Redistributions in binary form must reproduce the above copyright + * notice, this list of conditions and the following disclaimer in the + * documentation and/or other materials provided with the distribution. + * + * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" + * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE + * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE + * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE + * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR + * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF + * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS + * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN + * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) + * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE + * POSSIBILITY OF SUCH DAMAGE. + */ + +/* + * This file is part of the RTEMS quality process and was automatically + * generated. If you find something that needs to be fixed or + * worded better please post a report or patch to an RTEMS mailing list + * or raise a bug report: + * + * https://www.rtems.org/bugs.html + * + * For information on updating and regenerating please refer to the How-To + * section in the Software Requirements Engineering chapter of the + * RTEMS Software Engineering manual. The manual is provided as a part of + * a release. For development sources please refer to the online + * documentation at: + * + * https://docs.rtems.org + */ + +#ifndef _TR_TQ_FLUSH_PRIORITY_INHERIT_H +#define _TR_TQ_FLUSH_PRIORITY_INHERIT_H + +#include "tx-thread-queue.h" + +#ifdef __cplusplus +extern "C" { +#endif + +/** + * @addtogroup RTEMSTestCaseScoreTqReqFlushPriorityInherit + * + * @{ + */ + +typedef enum { + ScoreTqReqFlushPriorityInherit_Pre_Priority_Minimum, + ScoreTqReqFlushPriorityInherit_Pre_Priority_NotMinimum, + ScoreTqReqFlushPriorityInherit_Pre_Priority_NA +} ScoreTqReqFlushPriorityInherit_Pre_Priority; + +typedef enum { + ScoreTqReqFlushPriorityInherit_Pre_Queue_Empty, + ScoreTqReqFlushPriorityInherit_Pre_Queue_NonEmpty, + ScoreTqReqFlushPriorityInherit_Pre_Queue_NA +} ScoreTqReqFlushPriorityInherit_Pre_Queue; + +typedef enum { + ScoreTqReqFlushPriorityInherit_Post_Extract_Nop, + ScoreTqReqFlushPriorityInherit_Post_Extract_All, + ScoreTqReqFlushPriorityInherit_Post_Extract_NA +} ScoreTqReqFlushPriorityInherit_Post_Extract; + +typedef enum { + ScoreTqReqFlushPriorityInherit_Post_PriorityUpdate_No, + ScoreTqReqFlushPriorityInherit_Post_PriorityUpdate_Yes, + ScoreTqReqFlushPriorityInherit_Post_PriorityUpdate_NA +} ScoreTqReqFlushPriorityInherit_Post_PriorityUpdate; + +/** + * @brief Runs the parameterized test case. + * + * @param[in,out] tq_ctx is the thread queue test context. + */ +void ScoreTqReqFlushPriorityInherit_Run( TQContext *tq_ctx ); + +/** @} */ + +#ifdef __cplusplus +} +#endif + +#endif /* _TR_TQ_FLUSH_PRIORITY_INHERIT_H */ diff --git a/testsuites/validation/tr-tq-flush-priority.c b/testsuites/validation/tr-tq-flush-priority.c new file mode 100644 index 0000000000..10156cdf68 --- /dev/null +++ b/testsuites/validation/tr-tq-flush-priority.c @@ -0,0 +1,419 @@ +/* SPDX-License-Identifier: BSD-2-Clause */ + +/** + * @file + * + * @ingroup RTEMSTestCaseScoreTqReqFlushPriority + */ + +/* + * Copyright (C) 2021 embedded brains GmbH (http://www.embedded-brains.de) + * + * Redistribution and use in source and binary forms, with or without + * modification, are permitted provided that the following conditions + * are met: + * 1. Redistributions of source code must retain the above copyright + * notice, this list of conditions and the following disclaimer. + * 2. Redistributions in binary form must reproduce the above copyright + * notice, this list of conditions and the following disclaimer in the + * documentation and/or other materials provided with the distribution. + * + * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" + * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE + * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE + * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE + * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR + * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF + * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS + * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN + * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) + * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE + * POSSIBILITY OF SUCH DAMAGE. + */ + +/* + * This file is part of the RTEMS quality process and was automatically + * generated. If you find something that needs to be fixed or + * worded better please post a report or patch to an RTEMS mailing list + * or raise a bug report: + * + * https://www.rtems.org/bugs.html + * + * For information on updating and regenerating please refer to the How-To + * section in the Software Requirements Engineering chapter of the + * RTEMS Software Engineering manual. The manual is provided as a part of + * a release. For development sources please refer to the online + * documentation at: + * + * https://docs.rtems.org + */ + +#ifdef HAVE_CONFIG_H +#include "config.h" +#endif + +#include "tr-tq-flush-priority.h" +#include "tx-support.h" + +#include <rtems/test.h> + +/** + * @defgroup RTEMSTestCaseScoreTqReqFlushPriority \ + * spec:/score/tq/req/flush-priority + * + * @ingroup RTEMSTestSuiteTestsuitesValidation0 + * + * @{ + */ + +typedef struct { + uint8_t Skip : 1; + uint8_t Pre_Queue_NA : 1; + uint8_t Post_Operation : 2; +} ScoreTqReqFlushPriority_Entry; + +/** + * @brief Test context for spec:/score/tq/req/flush-priority test case. + */ +typedef struct { + /** + * @brief This member contains the call within ISR request. + */ + CallWithinISRRequest request;; + + /** + * @brief This member contains a copy of the corresponding + * ScoreTqReqFlushPriority_Run() parameter. + */ + TQContext *tq_ctx; + + /** + * @brief This member contains a copy of the corresponding + * ScoreTqReqFlushPriority_Run() parameter. + */ + bool supports_multiple_priority_queues; + + struct { + /** + * @brief This member defines the pre-condition states for the next action. + */ + size_t pcs[ 1 ]; + + /** + * @brief If this member is true, then the test action loop is executed. + */ + bool in_action_loop; + + /** + * @brief This member contains the next transition map index. + */ + size_t index; + + /** + * @brief This member contains the current transition map entry. + */ + ScoreTqReqFlushPriority_Entry entry; + + /** + * @brief If this member is true, then the current transition variant + * should be skipped. + */ + bool skip; + } Map; +} ScoreTqReqFlushPriority_Context; + +static ScoreTqReqFlushPriority_Context + ScoreTqReqFlushPriority_Instance; + +static const char * const ScoreTqReqFlushPriority_PreDesc_Queue[] = { + "Empty", + "NonEmpty", + "NA" +}; + +static const char * const * const ScoreTqReqFlushPriority_PreDesc[] = { + ScoreTqReqFlushPriority_PreDesc_Queue, + NULL +}; + +typedef ScoreTqReqFlushPriority_Context Context; + +static const T_scheduler_event *GetUnblock( Context *ctx, size_t *index ) +{ + return TQGetNextUnblock( ctx->tq_ctx, index ); +} + +static const rtems_tcb *GetTCB( Context *ctx, TQWorkerKind worker ) +{ + return ctx->tq_ctx->worker_tcb[ worker ]; +} + +static void Flush( void *arg ) +{ + Context *ctx; + + ctx = arg; + TQSchedulerRecordStart( ctx->tq_ctx ); + TQFlush( ctx->tq_ctx ); +} + +static void SchedulerEvent( + void *arg, + const T_scheduler_event *event, + T_scheduler_when when +) +{ + Context *ctx; + + ctx = arg; + + if ( + when == T_SCHEDULER_BEFORE && + event->operation == T_SCHEDULER_BLOCK + ) { + ctx->request.handler = Flush; + ctx->request.arg = ctx; + CallWithinISRSubmit( &ctx->request ); + T_scheduler_set_event_handler( NULL, NULL ); + } +} + +static void ScoreTqReqFlushPriority_Pre_Queue_Prepare( + ScoreTqReqFlushPriority_Context *ctx, + ScoreTqReqFlushPriority_Pre_Queue state +) +{ + switch ( state ) { + case ScoreTqReqFlushPriority_Pre_Queue_Empty: { + /* + * While the thread queue is empty. + */ + ctx->tq_ctx->how_many = 0; + break; + } + + case ScoreTqReqFlushPriority_Pre_Queue_NonEmpty: { + /* + * While the thread queue has at least one enqueued thread. + */ + ctx->tq_ctx->how_many = 3; + break; + } + + case ScoreTqReqFlushPriority_Pre_Queue_NA: + break; + } +} + +static void ScoreTqReqFlushPriority_Post_Operation_Check( + ScoreTqReqFlushPriority_Context *ctx, + ScoreTqReqFlushPriority_Post_Operation state +) +{ + size_t i; + const T_scheduler_event *event; + + i = 0; + + switch ( state ) { + case ScoreTqReqFlushPriority_Post_Operation_Nop: { + /* + * No operation shall be performed. + */ + /* Event receive */ + T_eq_ptr( GetUnblock( ctx, &i )->thread, GetTCB( ctx, TQ_BLOCKER_A ) ); + T_eq_ptr( GetUnblock( ctx, &i ), &T_scheduler_event_null ); + break; + } + + case ScoreTqReqFlushPriority_Post_Operation_TryExtract: { + /* + * The enqueued threads of the thread queue may be extracted in priority + * order for each priority queue associated with a scheduler. The + * priority queues of the thread queue shall be accessed in FIFO order. + */ + event = GetUnblock( ctx, &i ); + T_eq_ptr( event->executing, NULL ); + T_eq_ptr( event->thread, GetTCB( ctx, TQ_BLOCKER_D ) ); + + event = GetUnblock( ctx, &i ); + T_eq_ptr( event->executing, NULL ); + T_eq_ptr( event->thread, GetTCB( ctx, TQ_BLOCKER_C ) ); + + event = GetUnblock( ctx, &i ); + T_eq_ptr( event->executing, GetTCB( ctx, TQ_BLOCKER_B ) ); + T_eq_ptr( event->thread, GetTCB( ctx, TQ_BLOCKER_B ) ); + + T_eq_ptr( GetUnblock( ctx, &i ), &T_scheduler_event_null ); + break; + } + + case ScoreTqReqFlushPriority_Post_Operation_NA: + break; + } +} + +static void ScoreTqReqFlushPriority_Setup( + ScoreTqReqFlushPriority_Context *ctx +) +{ + TQReset( ctx->tq_ctx ); + TQSetPriority( ctx->tq_ctx, TQ_BLOCKER_A, PRIO_ULTRA_HIGH ); + + if ( + ctx->supports_multiple_priority_queues && + rtems_configuration_get_maximum_processors() > 1 + ) { + TQSetScheduler( ctx->tq_ctx, TQ_BLOCKER_B, SCHEDULER_B_ID, PRIO_HIGH ); + } else { + TQSetPriority( ctx->tq_ctx, TQ_BLOCKER_B, PRIO_HIGH ); + } + + TQSetPriority( ctx->tq_ctx, TQ_BLOCKER_C, PRIO_VERY_HIGH ); + TQSetPriority( ctx->tq_ctx, TQ_BLOCKER_D, PRIO_ULTRA_HIGH ); +} + +static void ScoreTqReqFlushPriority_Setup_Wrap( void *arg ) +{ + ScoreTqReqFlushPriority_Context *ctx; + + ctx = arg; + ctx->Map.in_action_loop = false; + ScoreTqReqFlushPriority_Setup( ctx ); +} + +static void ScoreTqReqFlushPriority_Teardown( + ScoreTqReqFlushPriority_Context *ctx +) +{ + TQReset( ctx->tq_ctx ); +} + +static void ScoreTqReqFlushPriority_Teardown_Wrap( void *arg ) +{ + ScoreTqReqFlushPriority_Context *ctx; + + ctx = arg; + ctx->Map.in_action_loop = false; + ScoreTqReqFlushPriority_Teardown( ctx ); +} + +static void ScoreTqReqFlushPriority_Action( + ScoreTqReqFlushPriority_Context *ctx +) +{ + TQSend( ctx->tq_ctx, TQ_BLOCKER_A, TQ_EVENT_ENQUEUE_PREPARE ); + + if ( ctx->tq_ctx->how_many > 0 ) { + TQSend( ctx->tq_ctx, TQ_BLOCKER_C, TQ_EVENT_ENQUEUE ); + TQSend( ctx->tq_ctx, TQ_BLOCKER_D, TQ_EVENT_ENQUEUE ); + T_scheduler_set_event_handler( SchedulerEvent, ctx ); + TQSendAndWaitForExecutionStop( + ctx->tq_ctx, + TQ_BLOCKER_B, + TQ_EVENT_ENQUEUE + ); + } else { + TQSchedulerRecordStart( ctx->tq_ctx ); + TQSend( ctx->tq_ctx, TQ_BLOCKER_A, TQ_EVENT_FLUSH ); + } + + TQSchedulerRecordStop( ctx->tq_ctx ); + TQSend( ctx->tq_ctx, TQ_BLOCKER_A, TQ_EVENT_ENQUEUE_DONE ); +} + +static const ScoreTqReqFlushPriority_Entry +ScoreTqReqFlushPriority_Entries[] = { + { 0, 0, ScoreTqReqFlushPriority_Post_Operation_Nop }, + { 0, 0, ScoreTqReqFlushPriority_Post_Operation_TryExtract } +}; + +static const uint8_t +ScoreTqReqFlushPriority_Map[] = { + 0, 1 +}; + +static size_t ScoreTqReqFlushPriority_Scope( void *arg, char *buf, size_t n ) +{ + ScoreTqReqFlushPriority_Context *ctx; + + ctx = arg; + + if ( ctx->Map.in_action_loop ) { + return T_get_scope( + ScoreTqReqFlushPriority_PreDesc, + buf, + n, + ctx->Map.pcs + ); + } + + return 0; +} + +static T_fixture ScoreTqReqFlushPriority_Fixture = { + .setup = ScoreTqReqFlushPriority_Setup_Wrap, + .stop = NULL, + .teardown = ScoreTqReqFlushPriority_Teardown_Wrap, + .scope = ScoreTqReqFlushPriority_Scope, + .initial_context = &ScoreTqReqFlushPriority_Instance +}; + +static inline ScoreTqReqFlushPriority_Entry ScoreTqReqFlushPriority_PopEntry( + ScoreTqReqFlushPriority_Context *ctx +) +{ + size_t index; + + index = ctx->Map.index; + ctx->Map.index = index + 1; + return ScoreTqReqFlushPriority_Entries[ + ScoreTqReqFlushPriority_Map[ index ] + ]; +} + +static void ScoreTqReqFlushPriority_TestVariant( + ScoreTqReqFlushPriority_Context *ctx +) +{ + ScoreTqReqFlushPriority_Pre_Queue_Prepare( ctx, ctx->Map.pcs[ 0 ] ); + ScoreTqReqFlushPriority_Action( ctx ); + ScoreTqReqFlushPriority_Post_Operation_Check( + ctx, + ctx->Map.entry.Post_Operation + ); +} + +static T_fixture_node ScoreTqReqFlushPriority_Node; + +void ScoreTqReqFlushPriority_Run( + TQContext *tq_ctx, + bool supports_multiple_priority_queues +) +{ + ScoreTqReqFlushPriority_Context *ctx; + + ctx = &ScoreTqReqFlushPriority_Instance; + ctx->tq_ctx = tq_ctx; + ctx->supports_multiple_priority_queues = supports_multiple_priority_queues; + + ctx = T_push_fixture( + &ScoreTqReqFlushPriority_Node, + &ScoreTqReqFlushPriority_Fixture + ); + ctx->Map.in_action_loop = true; + ctx->Map.index = 0; + + for ( + ctx->Map.pcs[ 0 ] = ScoreTqReqFlushPriority_Pre_Queue_Empty; + ctx->Map.pcs[ 0 ] < ScoreTqReqFlushPriority_Pre_Queue_NA; + ++ctx->Map.pcs[ 0 ] + ) { + ctx->Map.entry = ScoreTqReqFlushPriority_PopEntry( ctx ); + ScoreTqReqFlushPriority_TestVariant( ctx ); + } + + T_pop_fixture(); +} + +/** @} */ diff --git a/testsuites/validation/tr-tq-flush-priority.h b/testsuites/validation/tr-tq-flush-priority.h new file mode 100644 index 0000000000..72bc372ca1 --- /dev/null +++ b/testsuites/validation/tr-tq-flush-priority.h @@ -0,0 +1,97 @@ +/* SPDX-License-Identifier: BSD-2-Clause */ + +/** + * @file + * + * @ingroup RTEMSTestCaseScoreTqReqFlushPriority + */ + +/* + * Copyright (C) 2021 embedded brains GmbH (http://www.embedded-brains.de) + * + * Redistribution and use in source and binary forms, with or without + * modification, are permitted provided that the following conditions + * are met: + * 1. Redistributions of source code must retain the above copyright + * notice, this list of conditions and the following disclaimer. + * 2. Redistributions in binary form must reproduce the above copyright + * notice, this list of conditions and the following disclaimer in the + * documentation and/or other materials provided with the distribution. + * + * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" + * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE + * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE + * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE + * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR + * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF + * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS + * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN + * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) + * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE + * POSSIBILITY OF SUCH DAMAGE. + */ + +/* + * This file is part of the RTEMS quality process and was automatically + * generated. If you find something that needs to be fixed or + * worded better please post a report or patch to an RTEMS mailing list + * or raise a bug report: + * + * https://www.rtems.org/bugs.html + * + * For information on updating and regenerating please refer to the How-To + * section in the Software Requirements Engineering chapter of the + * RTEMS Software Engineering manual. The manual is provided as a part of + * a release. For development sources please refer to the online + * documentation at: + * + * https://docs.rtems.org + */ + +#ifndef _TR_TQ_FLUSH_PRIORITY_H +#define _TR_TQ_FLUSH_PRIORITY_H + +#include "tx-thread-queue.h" + +#ifdef __cplusplus +extern "C" { +#endif + +/** + * @addtogroup RTEMSTestCaseScoreTqReqFlushPriority + * + * @{ + */ + +typedef enum { + ScoreTqReqFlushPriority_Pre_Queue_Empty, + ScoreTqReqFlushPriority_Pre_Queue_NonEmpty, + ScoreTqReqFlushPriority_Pre_Queue_NA +} ScoreTqReqFlushPriority_Pre_Queue; + +typedef enum { + ScoreTqReqFlushPriority_Post_Operation_Nop, + ScoreTqReqFlushPriority_Post_Operation_TryExtract, + ScoreTqReqFlushPriority_Post_Operation_NA +} ScoreTqReqFlushPriority_Post_Operation; + +/** + * @brief Runs the parameterized test case. + * + * @param[in,out] tq_ctx is the thread queue test context. + * + * @param supports_multiple_priority_queues is true, if the object using the + * thread queue supports multiple priority queues, otherwise it is false. + */ +void ScoreTqReqFlushPriority_Run( + TQContext *tq_ctx, + bool supports_multiple_priority_queues +); + +/** @} */ + +#ifdef __cplusplus +} +#endif + +#endif /* _TR_TQ_FLUSH_PRIORITY_H */ |