/* SPDX-License-Identifier: BSD-2-Clause */ /** * @file * * @ingroup RTEMSTestCaseScoreMtxReqSeizeTry */ /* * Copyright (C) 2021 embedded brains GmbH (http://www.embedded-brains.de) * * Redistribution and use in source and binary forms, with or without * modification, are permitted provided that the following conditions * are met: * 1. Redistributions of source code must retain the above copyright * notice, this list of conditions and the following disclaimer. * 2. Redistributions in binary form must reproduce the above copyright * notice, this list of conditions and the following disclaimer in the * documentation and/or other materials provided with the distribution. * * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE * POSSIBILITY OF SUCH DAMAGE. */ /* * This file is part of the RTEMS quality process and was automatically * generated. If you find something that needs to be fixed or * worded better please post a report or patch to an RTEMS mailing list * or raise a bug report: * * https://www.rtems.org/bugs.html * * For information on updating and regenerating please refer to the How-To * section in the Software Requirements Engineering chapter of the * RTEMS Software Engineering manual. The manual is provided as a part of * a release. For development sources please refer to the online * documentation at: * * https://docs.rtems.org */ #ifdef HAVE_CONFIG_H #include "config.h" #endif #include "tr-mtx-seize-try.h" #include /** * @defgroup RTEMSTestCaseScoreMtxReqSeizeTry spec:/score/mtx/req/seize-try * * @ingroup RTEMSTestSuiteTestsuitesValidation0 * * @{ */ /** * @brief Test context for spec:/score/mtx/req/seize-try test case. */ typedef struct { /** * @brief If this member is true, then the calling thread shall be the owner * of the mutex. */ bool owner_self;; /** * @brief If this member is true, then a thread other than the calling thread * shall be the owner of the mutex. */ bool owner_other;; /** * @brief This member contains the current priority of the calling thread * before the directive call. */ rtems_task_priority priority_before;; /** * @brief This member contains the return status of the directive call. */ Status_Control status; /** * @brief This member contains the owner of the mutex after the directive * call. */ const rtems_tcb *owner_after; /** * @brief This member contains the current priority of the calling thread * after the directive call. */ rtems_task_priority priority_after;; /** * @brief This member contains a copy of the corresponding * ScoreMtxReqSeizeTry_Run() parameter. */ TQMtxContext *tq_ctx; /** * @brief This member defines the pre-condition states for the next action. */ size_t pcs[ 1 ]; /** * @brief This member indicates if the test action loop is currently * executed. */ bool in_action_loop; } ScoreMtxReqSeizeTry_Context; static ScoreMtxReqSeizeTry_Context ScoreMtxReqSeizeTry_Instance; static const char * const ScoreMtxReqSeizeTry_PreDesc_Owner[] = { "No", "Self", "Other", "NA" }; static const char * const * const ScoreMtxReqSeizeTry_PreDesc[] = { ScoreMtxReqSeizeTry_PreDesc_Owner, NULL }; typedef ScoreMtxReqSeizeTry_Context Context; static Status_Control Status( const Context *ctx, Status_Control status ) { return TQConvertStatus( &ctx->tq_ctx->base, status ); } static void ScoreMtxReqSeizeTry_Pre_Owner_Prepare( ScoreMtxReqSeizeTry_Context *ctx, ScoreMtxReqSeizeTry_Pre_Owner state ) { switch ( state ) { case ScoreMtxReqSeizeTry_Pre_Owner_No: { /* * While the mutex has no owner. */ /* This is the default */ break; } case ScoreMtxReqSeizeTry_Pre_Owner_Self: { /* * While the owner of the mutex is the calling thread. */ ctx->owner_self = true; break; } case ScoreMtxReqSeizeTry_Pre_Owner_Other: { /* * While the owner of the mutex is a thread other than the calling * thread. */ ctx->owner_other = true; break; } case ScoreMtxReqSeizeTry_Pre_Owner_NA: break; } } static void ScoreMtxReqSeizeTry_Post_Status_Check( ScoreMtxReqSeizeTry_Context *ctx, ScoreMtxReqSeizeTry_Post_Status state ) { switch ( state ) { case ScoreMtxReqSeizeTry_Post_Status_Ok: { /* * The return status of the directive call shall be derived from * STATUS_SUCCESSFUL. */ T_eq_int( ctx->status, Status( ctx, STATUS_SUCCESSFUL ) ); TQSurrender( &ctx->tq_ctx->base ); break; } case ScoreMtxReqSeizeTry_Post_Status_Recursive: { /* * Where the mutex supports a recursive seize, the return status of the * directive call shall be derived from STATUS_SUCCESSFUL. * * Where the mutex does not support a recursive seize, where a deadlock * is indicated by a status code, the return status of the directive call * shall be derived from STATUS_DEADLOCK. * * Where the mutex does not support a recursive seize, where a deadlock * is indicated by a fatal error, the thread queue deadlock internal * error shall occur. */ switch ( ctx->tq_ctx->recursive ) { case TQ_MTX_RECURSIVE_YES: T_eq_int( ctx->status, Status( ctx, STATUS_SUCCESSFUL ) ); TQSurrender( &ctx->tq_ctx->base ); break; case TQ_MTX_RECURSIVE_NO_STATUS: T_eq_int( ctx->status, Status( ctx, STATUS_DEADLOCK ) ); break; case TQ_MTX_RECURSIVE_NO_FATAL: /* TODO */ T_unreachable(); break; default: T_unreachable(); break; } break; } case ScoreMtxReqSeizeTry_Post_Status_Unsat: { /* * The return status of the directive call shall be derived from * STATUS_UNSATISFIED. */ T_eq_int( ctx->status, Status( ctx, STATUS_UNSATISFIED ) ); break; } case ScoreMtxReqSeizeTry_Post_Status_NA: break; } } static void ScoreMtxReqSeizeTry_Post_Owner_Check( ScoreMtxReqSeizeTry_Context *ctx, ScoreMtxReqSeizeTry_Post_Owner state ) { switch ( state ) { case ScoreMtxReqSeizeTry_Post_Owner_Nop: { /* * The owner of the semaphore shall not be modified. */ if ( ctx->owner_self ) { T_eq_ptr( ctx->owner_after, ctx->tq_ctx->base.runner_tcb ); } else if ( ctx->owner_other ) { T_eq_ptr( ctx->owner_after, ctx->tq_ctx->base.worker_tcb[ TQ_BLOCKER_A ] ); } else { T_null( ctx->owner_after ); } break; } case ScoreMtxReqSeizeTry_Post_Owner_New: { /* * The owner of the semaphore shall be the calling thread. */ T_eq_ptr( ctx->owner_after, ctx->tq_ctx->base.runner_tcb ); break; } case ScoreMtxReqSeizeTry_Post_Owner_NA: break; } } static void ScoreMtxReqSeizeTry_Post_Priority_Check( ScoreMtxReqSeizeTry_Context *ctx, ScoreMtxReqSeizeTry_Post_Priority state ) { switch ( state ) { case ScoreMtxReqSeizeTry_Post_Priority_Nop: { /* * The current priority of the calling thread shall not be modified. */ T_eq_u32( ctx->priority_after, ctx->priority_before ); break; } case ScoreMtxReqSeizeTry_Post_Priority_Ceiling: { /* * Where the mutex provides a priority ceiling, the calling thread shall * use the priority ceiling of the mutex. * * Where the mutex does not provide a priority ceiling, the current * priority of the calling thread shall not be modified. */ if ( ctx->tq_ctx->priority_ceiling != PRIO_INVALID ) { T_eq_u32( ctx->priority_after, ctx->tq_ctx->priority_ceiling ); } else { T_eq_u32( ctx->priority_after, ctx->priority_before ); } break; } case ScoreMtxReqSeizeTry_Post_Priority_NA: break; } } static void ScoreMtxReqSeizeTry_Prepare( ScoreMtxReqSeizeTry_Context *ctx ) { ctx->owner_self = false; ctx->owner_other = false; } static void ScoreMtxReqSeizeTry_Action( ScoreMtxReqSeizeTry_Context *ctx ) { if ( ctx->owner_self ) { Status_Control status; status = TQEnqueue( &ctx->tq_ctx->base, TQ_NO_WAIT ); T_eq_int( status, Status( ctx, STATUS_SUCCESSFUL ) ); } else if ( ctx->owner_other ) { if ( ctx->tq_ctx->base.enqueue_variant == TQ_ENQUEUE_STICKY ) { TQSetScheduler( &ctx->tq_ctx->base, TQ_BLOCKER_A, ctx->tq_ctx->base.other_scheduler_id, PRIO_HIGH ); TQSend( &ctx->tq_ctx->base, TQ_BLOCKER_A, TQ_EVENT_ENQUEUE | TQ_EVENT_RUNNER_SYNC ); TQSynchronizeRunner(); } else { TQSend( &ctx->tq_ctx->base, TQ_BLOCKER_A, TQ_EVENT_ENQUEUE ); } } ctx->priority_before = GetSelfPriority(); ctx->status = TQEnqueue( &ctx->tq_ctx->base, TQ_NO_WAIT ); ctx->owner_after = TQMtxGetOwner( ctx->tq_ctx ); ctx->priority_after = GetSelfPriority(); if ( ctx->owner_self ) { TQSurrender( &ctx->tq_ctx->base ); } else if ( ctx->owner_other ) { if ( ctx->tq_ctx->base.enqueue_variant == TQ_ENQUEUE_STICKY ) { TQSend( &ctx->tq_ctx->base, TQ_BLOCKER_A, TQ_EVENT_SURRENDER | TQ_EVENT_RUNNER_SYNC ); TQSynchronizeRunner(); TQSetScheduler( &ctx->tq_ctx->base, TQ_BLOCKER_A, ctx->tq_ctx->base.runner_scheduler_id, PRIO_HIGH ); } else { TQSend( &ctx->tq_ctx->base, TQ_BLOCKER_A, TQ_EVENT_SURRENDER ); } } } typedef struct { uint8_t Skip : 1; uint8_t Pre_Owner_NA : 1; uint8_t Post_Status : 2; uint8_t Post_Owner : 2; uint8_t Post_Priority : 2; } ScoreMtxReqSeizeTry_Entry; static const ScoreMtxReqSeizeTry_Entry ScoreMtxReqSeizeTry_Entries[] = { { 0, 0, ScoreMtxReqSeizeTry_Post_Status_Ok, ScoreMtxReqSeizeTry_Post_Owner_New, ScoreMtxReqSeizeTry_Post_Priority_Ceiling }, { 0, 0, ScoreMtxReqSeizeTry_Post_Status_Recursive, ScoreMtxReqSeizeTry_Post_Owner_Nop, ScoreMtxReqSeizeTry_Post_Priority_Nop }, { 0, 0, ScoreMtxReqSeizeTry_Post_Status_Unsat, ScoreMtxReqSeizeTry_Post_Owner_Nop, ScoreMtxReqSeizeTry_Post_Priority_Nop } }; static const uint8_t ScoreMtxReqSeizeTry_Map[] = { 0, 1, 2 }; static size_t ScoreMtxReqSeizeTry_Scope( void *arg, char *buf, size_t n ) { ScoreMtxReqSeizeTry_Context *ctx; ctx = arg; if ( ctx->in_action_loop ) { return T_get_scope( ScoreMtxReqSeizeTry_PreDesc, buf, n, ctx->pcs ); } return 0; } static T_fixture ScoreMtxReqSeizeTry_Fixture = { .setup = NULL, .stop = NULL, .teardown = NULL, .scope = ScoreMtxReqSeizeTry_Scope, .initial_context = &ScoreMtxReqSeizeTry_Instance }; static inline ScoreMtxReqSeizeTry_Entry ScoreMtxReqSeizeTry_GetEntry( size_t index ) { return ScoreMtxReqSeizeTry_Entries[ ScoreMtxReqSeizeTry_Map[ index ] ]; } static T_fixture_node ScoreMtxReqSeizeTry_Node; void ScoreMtxReqSeizeTry_Run( TQMtxContext *tq_ctx ) { ScoreMtxReqSeizeTry_Context *ctx; size_t index; ctx = &ScoreMtxReqSeizeTry_Instance; ctx->tq_ctx = tq_ctx; ctx = T_push_fixture( &ScoreMtxReqSeizeTry_Node, &ScoreMtxReqSeizeTry_Fixture ); ctx->in_action_loop = true; index = 0; for ( ctx->pcs[ 0 ] = ScoreMtxReqSeizeTry_Pre_Owner_No; ctx->pcs[ 0 ] < ScoreMtxReqSeizeTry_Pre_Owner_NA; ++ctx->pcs[ 0 ] ) { ScoreMtxReqSeizeTry_Entry entry; entry = ScoreMtxReqSeizeTry_GetEntry( index ); ++index; ScoreMtxReqSeizeTry_Prepare( ctx ); ScoreMtxReqSeizeTry_Pre_Owner_Prepare( ctx, ctx->pcs[ 0 ] ); ScoreMtxReqSeizeTry_Action( ctx ); ScoreMtxReqSeizeTry_Post_Status_Check( ctx, entry.Post_Status ); ScoreMtxReqSeizeTry_Post_Owner_Check( ctx, entry.Post_Owner ); ScoreMtxReqSeizeTry_Post_Priority_Check( ctx, entry.Post_Priority ); } T_pop_fixture(); } /** @} */