/* SPDX-License-Identifier: BSD-2-Clause */
/**
* @file
*
* @ingroup RTEMSTestCaseScoreTqReqFlushPriority
*/
/*
* Copyright (C) 2021 embedded brains GmbH (http://www.embedded-brains.de)
*
* Redistribution and use in source and binary forms, with or without
* modification, are permitted provided that the following conditions
* are met:
* 1. Redistributions of source code must retain the above copyright
* notice, this list of conditions and the following disclaimer.
* 2. Redistributions in binary form must reproduce the above copyright
* notice, this list of conditions and the following disclaimer in the
* documentation and/or other materials provided with the distribution.
*
* THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
* AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
* IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
* ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE
* LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
* CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
* SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
* INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
* CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
* ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
* POSSIBILITY OF SUCH DAMAGE.
*/
/*
* This file is part of the RTEMS quality process and was automatically
* generated. If you find something that needs to be fixed or
* worded better please post a report or patch to an RTEMS mailing list
* or raise a bug report:
*
* https://www.rtems.org/bugs.html
*
* For information on updating and regenerating please refer to the How-To
* section in the Software Requirements Engineering chapter of the
* RTEMS Software Engineering manual. The manual is provided as a part of
* a release. For development sources please refer to the online
* documentation at:
*
* https://docs.rtems.org
*/
#ifdef HAVE_CONFIG_H
#include "config.h"
#endif
#include "tr-tq-flush-priority.h"
#include "tx-support.h"
#include <rtems/test.h>
/**
* @defgroup RTEMSTestCaseScoreTqReqFlushPriority \
* spec:/score/tq/req/flush-priority
*
* @ingroup RTEMSTestSuiteTestsuitesValidationNoClock0
*
* @{
*/
typedef struct {
uint8_t Skip : 1;
uint8_t Pre_Queue_NA : 1;
uint8_t Post_Operation : 2;
} ScoreTqReqFlushPriority_Entry;
/**
* @brief Test context for spec:/score/tq/req/flush-priority test case.
*/
typedef struct {
/**
* @brief This member contains the call within ISR request.
*/
CallWithinISRRequest request;;
/**
* @brief This member contains a copy of the corresponding
* ScoreTqReqFlushPriority_Run() parameter.
*/
TQContext *tq_ctx;
/**
* @brief This member contains a copy of the corresponding
* ScoreTqReqFlushPriority_Run() parameter.
*/
bool supports_multiple_priority_queues;
struct {
/**
* @brief This member defines the pre-condition states for the next action.
*/
size_t pcs[ 1 ];
/**
* @brief If this member is true, then the test action loop is executed.
*/
bool in_action_loop;
/**
* @brief This member contains the next transition map index.
*/
size_t index;
/**
* @brief This member contains the current transition map entry.
*/
ScoreTqReqFlushPriority_Entry entry;
/**
* @brief If this member is true, then the current transition variant
* should be skipped.
*/
bool skip;
} Map;
} ScoreTqReqFlushPriority_Context;
static ScoreTqReqFlushPriority_Context
ScoreTqReqFlushPriority_Instance;
static const char * const ScoreTqReqFlushPriority_PreDesc_Queue[] = {
"Empty",
"NonEmpty",
"NA"
};
static const char * const * const ScoreTqReqFlushPriority_PreDesc[] = {
ScoreTqReqFlushPriority_PreDesc_Queue,
NULL
};
typedef ScoreTqReqFlushPriority_Context Context;
static const T_scheduler_event *GetUnblock( Context *ctx, size_t *index )
{
return TQGetNextUnblock( ctx->tq_ctx, index );
}
static const rtems_tcb *GetTCB( Context *ctx, TQWorkerKind worker )
{
return ctx->tq_ctx->worker_tcb[ worker ];
}
static void Flush( void *arg )
{
Context *ctx;
ctx = arg;
TQSchedulerRecordStart( ctx->tq_ctx );
TQFlush( ctx->tq_ctx );
}
static void SchedulerEvent(
void *arg,
const T_scheduler_event *event,
T_scheduler_when when
)
{
Context *ctx;
ctx = arg;
if (
when == T_SCHEDULER_BEFORE &&
event->operation == T_SCHEDULER_BLOCK
) {
ctx->request.handler = Flush;
ctx->request.arg = ctx;
CallWithinISRSubmit( &ctx->request );
T_scheduler_set_event_handler( NULL, NULL );
}
}
static void ScoreTqReqFlushPriority_Pre_Queue_Prepare(
ScoreTqReqFlushPriority_Context *ctx,
ScoreTqReqFlushPriority_Pre_Queue state
)
{
switch ( state ) {
case ScoreTqReqFlushPriority_Pre_Queue_Empty: {
/*
* While the thread queue is empty.
*/
ctx->tq_ctx->how_many = 0;
break;
}
case ScoreTqReqFlushPriority_Pre_Queue_NonEmpty: {
/*
* While the thread queue has at least one enqueued thread.
*/
ctx->tq_ctx->how_many = 3;
break;
}
case ScoreTqReqFlushPriority_Pre_Queue_NA:
break;
}
}
static void ScoreTqReqFlushPriority_Post_Operation_Check(
ScoreTqReqFlushPriority_Context *ctx,
ScoreTqReqFlushPriority_Post_Operation state
)
{
size_t i;
const T_scheduler_event *event;
i = 0;
switch ( state ) {
case ScoreTqReqFlushPriority_Post_Operation_Nop: {
/*
* No operation shall be performed.
*/
/* Event receive */
T_eq_ptr( GetUnblock( ctx, &i )->thread, GetTCB( ctx, TQ_BLOCKER_A ) );
T_eq_ptr( GetUnblock( ctx, &i ), &T_scheduler_event_null );
break;
}
case ScoreTqReqFlushPriority_Post_Operation_TryExtract: {
/*
* The enqueued threads of the thread queue may be extracted in priority
* order for each priority queue associated with a scheduler. The
* priority queues of the thread queue shall be accessed in FIFO order.
*/
event = GetUnblock( ctx, &i );
T_eq_ptr( event->executing, NULL );
T_eq_ptr( event->thread, GetTCB( ctx, TQ_BLOCKER_D ) );
event = GetUnblock( ctx, &i );
T_eq_ptr( event->executing, NULL );
T_eq_ptr( event->thread, GetTCB( ctx, TQ_BLOCKER_C ) );
event = GetUnblock( ctx, &i );
T_eq_ptr( event->executing, GetTCB( ctx, TQ_BLOCKER_B ) );
T_eq_ptr( event->thread, GetTCB( ctx, TQ_BLOCKER_B ) );
T_eq_ptr( GetUnblock( ctx, &i ), &T_scheduler_event_null );
break;
}
case ScoreTqReqFlushPriority_Post_Operation_NA:
break;
}
}
static void ScoreTqReqFlushPriority_Setup(
ScoreTqReqFlushPriority_Context *ctx
)
{
TQReset( ctx->tq_ctx );
TQSetPriority( ctx->tq_ctx, TQ_BLOCKER_A, PRIO_ULTRA_HIGH );
if (
ctx->supports_multiple_priority_queues &&
rtems_configuration_get_maximum_processors() > 1
) {
TQSetScheduler( ctx->tq_ctx, TQ_BLOCKER_B, SCHEDULER_B_ID, PRIO_HIGH );
} else {
TQSetPriority( ctx->tq_ctx, TQ_BLOCKER_B, PRIO_HIGH );
}
TQSetPriority( ctx->tq_ctx, TQ_BLOCKER_C, PRIO_VERY_HIGH );
TQSetPriority( ctx->tq_ctx, TQ_BLOCKER_D, PRIO_ULTRA_HIGH );
}
static void ScoreTqReqFlushPriority_Setup_Wrap( void *arg )
{
ScoreTqReqFlushPriority_Context *ctx;
ctx = arg;
ctx->Map.in_action_loop = false;
ScoreTqReqFlushPriority_Setup( ctx );
}
static void ScoreTqReqFlushPriority_Teardown(
ScoreTqReqFlushPriority_Context *ctx
)
{
TQReset( ctx->tq_ctx );
}
static void ScoreTqReqFlushPriority_Teardown_Wrap( void *arg )
{
ScoreTqReqFlushPriority_Context *ctx;
ctx = arg;
ctx->Map.in_action_loop = false;
ScoreTqReqFlushPriority_Teardown( ctx );
}
static void ScoreTqReqFlushPriority_Action(
ScoreTqReqFlushPriority_Context *ctx
)
{
TQSend( ctx->tq_ctx, TQ_BLOCKER_A, TQ_EVENT_ENQUEUE_PREPARE );
if ( ctx->tq_ctx->how_many > 0 ) {
TQSend( ctx->tq_ctx, TQ_BLOCKER_C, TQ_EVENT_ENQUEUE );
TQSend( ctx->tq_ctx, TQ_BLOCKER_D, TQ_EVENT_ENQUEUE );
T_scheduler_set_event_handler( SchedulerEvent, ctx );
TQSendAndWaitForExecutionStop(
ctx->tq_ctx,
TQ_BLOCKER_B,
TQ_EVENT_ENQUEUE
);
} else {
TQSchedulerRecordStart( ctx->tq_ctx );
TQSend( ctx->tq_ctx, TQ_BLOCKER_A, TQ_EVENT_FLUSH );
}
TQSchedulerRecordStop( ctx->tq_ctx );
TQSend( ctx->tq_ctx, TQ_BLOCKER_A, TQ_EVENT_ENQUEUE_DONE );
}
static const ScoreTqReqFlushPriority_Entry
ScoreTqReqFlushPriority_Entries[] = {
{ 0, 0, ScoreTqReqFlushPriority_Post_Operation_Nop },
{ 0, 0, ScoreTqReqFlushPriority_Post_Operation_TryExtract }
};
static const uint8_t
ScoreTqReqFlushPriority_Map[] = {
0, 1
};
static size_t ScoreTqReqFlushPriority_Scope( void *arg, char *buf, size_t n )
{
ScoreTqReqFlushPriority_Context *ctx;
ctx = arg;
if ( ctx->Map.in_action_loop ) {
return T_get_scope(
ScoreTqReqFlushPriority_PreDesc,
buf,
n,
ctx->Map.pcs
);
}
return 0;
}
static T_fixture ScoreTqReqFlushPriority_Fixture = {
.setup = ScoreTqReqFlushPriority_Setup_Wrap,
.stop = NULL,
.teardown = ScoreTqReqFlushPriority_Teardown_Wrap,
.scope = ScoreTqReqFlushPriority_Scope,
.initial_context = &ScoreTqReqFlushPriority_Instance
};
static inline ScoreTqReqFlushPriority_Entry ScoreTqReqFlushPriority_PopEntry(
ScoreTqReqFlushPriority_Context *ctx
)
{
size_t index;
index = ctx->Map.index;
ctx->Map.index = index + 1;
return ScoreTqReqFlushPriority_Entries[
ScoreTqReqFlushPriority_Map[ index ]
];
}
static void ScoreTqReqFlushPriority_TestVariant(
ScoreTqReqFlushPriority_Context *ctx
)
{
ScoreTqReqFlushPriority_Pre_Queue_Prepare( ctx, ctx->Map.pcs[ 0 ] );
ScoreTqReqFlushPriority_Action( ctx );
ScoreTqReqFlushPriority_Post_Operation_Check(
ctx,
ctx->Map.entry.Post_Operation
);
}
static T_fixture_node ScoreTqReqFlushPriority_Node;
void ScoreTqReqFlushPriority_Run(
TQContext *tq_ctx,
bool supports_multiple_priority_queues
)
{
ScoreTqReqFlushPriority_Context *ctx;
ctx = &ScoreTqReqFlushPriority_Instance;
ctx->tq_ctx = tq_ctx;
ctx->supports_multiple_priority_queues = supports_multiple_priority_queues;
ctx = T_push_fixture(
&ScoreTqReqFlushPriority_Node,
&ScoreTqReqFlushPriority_Fixture
);
ctx->Map.in_action_loop = true;
ctx->Map.index = 0;
for (
ctx->Map.pcs[ 0 ] = ScoreTqReqFlushPriority_Pre_Queue_Empty;
ctx->Map.pcs[ 0 ] < ScoreTqReqFlushPriority_Pre_Queue_NA;
++ctx->Map.pcs[ 0 ]
) {
ctx->Map.entry = ScoreTqReqFlushPriority_PopEntry( ctx );
ScoreTqReqFlushPriority_TestVariant( ctx );
}
T_pop_fixture();
}
/** @} */