diff options
author | Sebastian Huber <sebastian.huber@embedded-brains.de> | 2021-11-11 08:03:38 +0100 |
---|---|---|
committer | Sebastian Huber <sebastian.huber@embedded-brains.de> | 2021-11-11 17:05:10 +0100 |
commit | 6d72776ca832b8efd12a133da94aae09062667f2 (patch) | |
tree | 475c8a8525f282d0d7d8b7a634f191838b464697 | |
parent | eb400f06e69e08de47b0a4abd772018ba0ddd008 (diff) |
testsuites/validation/tc-scheduler-remove-processor.c
-rw-r--r-- | testsuites/validation/tc-scheduler-remove-processor.c | 1116 |
1 files changed, 979 insertions, 137 deletions
diff --git a/testsuites/validation/tc-scheduler-remove-processor.c b/testsuites/validation/tc-scheduler-remove-processor.c index 4815353c77..e9967cb35a 100644 --- a/testsuites/validation/tc-scheduler-remove-processor.c +++ b/testsuites/validation/tc-scheduler-remove-processor.c @@ -54,6 +54,8 @@ #include <rtems.h> #include <rtems/test-scheduler.h> +#include <rtems/score/percpu.h> +#include <rtems/score/smpbarrier.h> #include "ts-config.h" #include "tx-support.h" @@ -70,13 +72,6 @@ */ typedef enum { - RtemsSchedulerReqRemoveProcessor_Pre_CPUState_Idle, - RtemsSchedulerReqRemoveProcessor_Pre_CPUState_InUse, - RtemsSchedulerReqRemoveProcessor_Pre_CPUState_NotOwned, - RtemsSchedulerReqRemoveProcessor_Pre_CPUState_NA -} RtemsSchedulerReqRemoveProcessor_Pre_CPUState; - -typedef enum { RtemsSchedulerReqRemoveProcessor_Pre_Id_Invalid, RtemsSchedulerReqRemoveProcessor_Pre_Id_Scheduler, RtemsSchedulerReqRemoveProcessor_Pre_Id_NA @@ -89,6 +84,38 @@ typedef enum { } RtemsSchedulerReqRemoveProcessor_Pre_CPUIndex; typedef enum { + RtemsSchedulerReqRemoveProcessor_Pre_Owned_Yes, + RtemsSchedulerReqRemoveProcessor_Pre_Owned_No, + RtemsSchedulerReqRemoveProcessor_Pre_Owned_NA +} RtemsSchedulerReqRemoveProcessor_Pre_Owned; + +typedef enum { + RtemsSchedulerReqRemoveProcessor_Pre_Last_Yes, + RtemsSchedulerReqRemoveProcessor_Pre_Last_No, + RtemsSchedulerReqRemoveProcessor_Pre_Last_NA +} RtemsSchedulerReqRemoveProcessor_Pre_Last; + +typedef enum { + RtemsSchedulerReqRemoveProcessor_Pre_Home_Yes, + RtemsSchedulerReqRemoveProcessor_Pre_Home_No, + RtemsSchedulerReqRemoveProcessor_Pre_Home_NA +} RtemsSchedulerReqRemoveProcessor_Pre_Home; + +typedef enum { + RtemsSchedulerReqRemoveProcessor_Pre_RequiredByAffinity_Yes, + RtemsSchedulerReqRemoveProcessor_Pre_RequiredByAffinity_No, + RtemsSchedulerReqRemoveProcessor_Pre_RequiredByAffinity_NA +} RtemsSchedulerReqRemoveProcessor_Pre_RequiredByAffinity; + +typedef enum { + RtemsSchedulerReqRemoveProcessor_Pre_UsedBy_Idle, + RtemsSchedulerReqRemoveProcessor_Pre_UsedBy_Task, + RtemsSchedulerReqRemoveProcessor_Pre_UsedBy_TaskIdle, + RtemsSchedulerReqRemoveProcessor_Pre_UsedBy_Helping, + RtemsSchedulerReqRemoveProcessor_Pre_UsedBy_NA +} RtemsSchedulerReqRemoveProcessor_Pre_UsedBy; + +typedef enum { RtemsSchedulerReqRemoveProcessor_Post_Status_Ok, RtemsSchedulerReqRemoveProcessor_Post_Status_InvId, RtemsSchedulerReqRemoveProcessor_Post_Status_InvNum, @@ -104,42 +131,122 @@ typedef enum { typedef struct { uint16_t Skip : 1; - uint16_t Pre_CPUState_NA : 1; uint16_t Pre_Id_NA : 1; uint16_t Pre_CPUIndex_NA : 1; + uint16_t Pre_Owned_NA : 1; + uint16_t Pre_Last_NA : 1; + uint16_t Pre_Home_NA : 1; + uint16_t Pre_RequiredByAffinity_NA : 1; + uint16_t Pre_UsedBy_NA : 1; uint16_t Post_Status : 3; uint16_t Post_Removed : 2; } RtemsSchedulerReqRemoveProcessor_Entry; +typedef enum { + WORKER_A, + WORKER_B, + WORKER_C, + WORKER_COUNT +} WorkerIndex; + /** * @brief Test context for spec:/rtems/scheduler/req/remove-processor test * case. */ typedef struct { /** - * @brief This member specifies the scheduler used to add the processor. + * @brief This member contains the runner identifier. + */ + rtems_id runner_id; + + /** + * @brief This member contains the worker identifiers. + */ + rtems_id worker_id[ WORKER_COUNT ]; + + /** + * @brief This member contains the mutex identifier. + */ + rtems_id mutex_id; + + /** + * @brief This member contains the sticky mutex identifier. + */ + rtems_id sticky_id; + + /** + * @brief This member contains the worker busy status. */ - rtems_id scheduler_id; + volatile bool busy[ WORKER_COUNT ];; /** - * @brief This member contains the identifier of scheduler A. + * @brief This member contains the worker busy status. */ - rtems_id scheduler_a_id; + volatile uint32_t busy_counter[ WORKER_COUNT ];; /** - * @brief This member contains the identifier of scheduler B. + * @brief This member contains the barrier to synchronize the runner and the + * workers. */ - rtems_id scheduler_b_id; + SMP_barrier_Control barrier; /** - * @brief This member specifies the processor to remove. + * @brief This member contains the call within ISR request. */ - uint32_t cpu_to_remove; + CallWithinISRRequest request;; + + /** + * @brief This member provides the context to wrap thread queue operations. + */ + WrapThreadQueueContext wrap_tq_ctx; + + /** + * @brief If this member is true, then the processor to remove shall be owned + * by the scheduler. + */ + bool owned; + + /** + * @brief If this member is true, then the processor to remove shall be the + * last processor of the scheduler. + */ + bool last; + + /** + * @brief If this member is true, then at least one non-idle task shall use + * the scheduler as its home scheduler. + */ + bool home; + + /** + * @brief If this member is true, then at least one non-idle task shall + * required the processor to remove due to its affinity set. + */ + bool required_by_affinity; + + /** + * @brief If this member is true, then the processor to remove shall be used + * by an idle task. + */ + bool idle; + + /** + * @brief If this member is true, then the processor to remove shall be used + * by a task or on behalf of a task which uses the scheduler as its home + * scheduler. + */ + bool task; + + /** + * @brief If this member is true, then the processor to remove shall be used + * by a task which uses the scheduler as a helping scheduler. + */ + bool helping; /** * @brief This member provides the scheduler operation records. */ - T_scheduler_log_2 scheduler_log;; + T_scheduler_log_10 scheduler_log;; /** * @brief This member contains the return value of the @@ -148,12 +255,12 @@ typedef struct { rtems_status_code status; /** - * @brief This member specifies if the ``scheduler_id`` parameter value. + * @brief This member specifies the ``scheduler_id`` parameter value. */ rtems_id id; /** - * @brief This member specifies if the ``cpu_index`` parameter value. + * @brief This member specifies the ``cpu_index`` parameter value. */ uint32_t cpu_index; @@ -162,12 +269,12 @@ typedef struct { * @brief This member defines the pre-condition indices for the next * action. */ - size_t pci[ 3 ]; + size_t pci[ 7 ]; /** * @brief This member defines the pre-condition states for the next action. */ - size_t pcs[ 3 ]; + size_t pcs[ 7 ]; /** * @brief If this member is true, then the test action loop is executed. @@ -195,13 +302,6 @@ typedef struct { static RtemsSchedulerReqRemoveProcessor_Context RtemsSchedulerReqRemoveProcessor_Instance; -static const char * const RtemsSchedulerReqRemoveProcessor_PreDesc_CPUState[] = { - "Idle", - "InUse", - "NotOwned", - "NA" -}; - static const char * const RtemsSchedulerReqRemoveProcessor_PreDesc_Id[] = { "Invalid", "Scheduler", @@ -214,62 +314,281 @@ static const char * const RtemsSchedulerReqRemoveProcessor_PreDesc_CPUIndex[] = "NA" }; +static const char * const RtemsSchedulerReqRemoveProcessor_PreDesc_Owned[] = { + "Yes", + "No", + "NA" +}; + +static const char * const RtemsSchedulerReqRemoveProcessor_PreDesc_Last[] = { + "Yes", + "No", + "NA" +}; + +static const char * const RtemsSchedulerReqRemoveProcessor_PreDesc_Home[] = { + "Yes", + "No", + "NA" +}; + +static const char * const RtemsSchedulerReqRemoveProcessor_PreDesc_RequiredByAffinity[] = { + "Yes", + "No", + "NA" +}; + +static const char * const RtemsSchedulerReqRemoveProcessor_PreDesc_UsedBy[] = { + "Idle", + "Task", + "TaskIdle", + "Helping", + "NA" +}; + static const char * const * const RtemsSchedulerReqRemoveProcessor_PreDesc[] = { - RtemsSchedulerReqRemoveProcessor_PreDesc_CPUState, RtemsSchedulerReqRemoveProcessor_PreDesc_Id, RtemsSchedulerReqRemoveProcessor_PreDesc_CPUIndex, + RtemsSchedulerReqRemoveProcessor_PreDesc_Owned, + RtemsSchedulerReqRemoveProcessor_PreDesc_Last, + RtemsSchedulerReqRemoveProcessor_PreDesc_Home, + RtemsSchedulerReqRemoveProcessor_PreDesc_RequiredByAffinity, + RtemsSchedulerReqRemoveProcessor_PreDesc_UsedBy, NULL }; -static void RtemsSchedulerReqRemoveProcessor_Pre_CPUState_Prepare( - RtemsSchedulerReqRemoveProcessor_Context *ctx, - RtemsSchedulerReqRemoveProcessor_Pre_CPUState state -) +typedef RtemsSchedulerReqRemoveProcessor_Context Context; + +static void DoRemoveProcessor( Context *ctx ) { - ctx->scheduler_id = ctx->scheduler_a_id; - ctx->cpu_to_remove = 0; + T_scheduler_log *log; - switch ( state ) { - case RtemsSchedulerReqRemoveProcessor_Pre_CPUState_Idle: { - /* - * While the processor associated with the ``cpu_index`` parameter is - * owned by the scheduler specified by the ``scheduler_id`` parameter, - * while no task exists which uses the scheduler as its home scheduler - * and the affinity set of this task would require the processor - * specified by the ``cpu_index`` parameter. - */ - ctx->scheduler_id = ctx->scheduler_b_id; - ctx->cpu_to_remove = 1; - break; + log = T_scheduler_record_10( &ctx->scheduler_log ); + T_null( log ); + + ctx->status = rtems_scheduler_remove_processor( ctx->id, ctx->cpu_index ); + + log = T_scheduler_record( NULL ); + T_eq_ptr( &log->header, &ctx->scheduler_log.header ); + + if ( ctx->status == RTEMS_SUCCESSFUL ) { + AddProcessor( ctx->id, ctx->cpu_index ); + } +} + +#if defined(RTEMS_SMP) +typedef enum { + EVENT_SYNC_RUNNER = RTEMS_EVENT_0, + EVENT_OBTAIN = RTEMS_EVENT_1, + EVENT_RELEASE = RTEMS_EVENT_2, + EVENT_STICKY_OBTAIN = RTEMS_EVENT_3, + EVENT_STICKY_RELEASE = RTEMS_EVENT_4, + EVENT_RESTART = RTEMS_EVENT_5, + EVENT_BUSY = RTEMS_EVENT_6, + EVENT_SYNC_RUNNER_LATE = RTEMS_EVENT_7 +} Event; + +static void Barriers( void *arg ) +{ + Context *ctx; + SMP_barrier_State barrier_state; + + ctx = arg; + _SMP_barrier_State_initialize( &barrier_state ); + + /* A */ + _SMP_barrier_Wait( &ctx->barrier, &barrier_state, 2 ); + + /* B */ + _SMP_barrier_Wait( &ctx->barrier, &barrier_state, 2 ); +} + +static void RequestISR( void *arg ) +{ + Context *ctx; + + ctx = arg; + ctx->request.handler = Barriers; + ctx->request.arg = ctx; + CallWithinISRSubmit( &ctx->request ); +} + +static void SendAndSync( Context *ctx, WorkerIndex worker, Event event ) +{ + SendEvents( ctx->worker_id[ worker ], EVENT_SYNC_RUNNER | event ); + ReceiveAllEvents( EVENT_SYNC_RUNNER ); + WaitForExecutionStop( ctx->worker_id[ worker ] ); +} + +static void MakeBusy( Context *ctx, WorkerIndex worker ) +{ + ctx->busy_counter[ worker ] = 0; + ctx->busy[ worker ] = true; + SendEvents( ctx->worker_id[ worker ], EVENT_BUSY ); +} + +static void MakeBusyAndSync( Context *ctx, WorkerIndex worker ) +{ + ctx->busy_counter[ worker ] = 0; + ctx->busy[ worker ] = true; + SendEvents( ctx->worker_id[ worker ], EVENT_SYNC_RUNNER | EVENT_BUSY ); + ReceiveAllEvents( EVENT_SYNC_RUNNER ); +} + +static void StopBusy( Context *ctx, WorkerIndex worker ) +{ + ctx->busy[ worker ] = false; +} + +static void StopBusyAndWait( Context *ctx, WorkerIndex worker ) +{ + StopBusy( ctx, worker ); + WaitForExecutionStop( ctx->worker_id[ worker ] ); +} + +static void WaitForBusy( Context *ctx, WorkerIndex worker ) +{ + while ( ctx->busy_counter[ worker ] == 0 ) { + /* Wait */ + } +} + +static void RemoveWithHelpingOnly( Context *ctx ) +{ + SMP_barrier_State barrier_state; + + /* + * Use the mutex and the worker to construct the removal of the last + * processor of a scheduler while a thread is scheduled. + */ + + _SMP_barrier_Control_initialize( &ctx->barrier ); + _SMP_barrier_State_initialize( &barrier_state ); + + SetScheduler( ctx->worker_id[ WORKER_B ], SCHEDULER_B_ID, PRIO_NORMAL ); + + /* Let worker B help worker A */ + SendEvents( ctx->worker_id[ WORKER_A ], EVENT_OBTAIN ); + SendAndSync( ctx, WORKER_B, EVENT_OBTAIN ); + + /* + * Restart the worker B to withdraw the help offer and wait on barriers. + * Move worker B to scheduler A. Remove the processor while worker A is + * scheduled. + */ + + SendEvents( ctx->worker_id[ WORKER_A ], EVENT_RESTART ); + + /* A */ + _SMP_barrier_Wait( &ctx->barrier, &barrier_state, 2 ); + + SetScheduler( ctx->worker_id[ WORKER_B ], SCHEDULER_A_ID, PRIO_HIGH ); + + ctx->id = SCHEDULER_B_ID; + ctx->cpu_index = 1; + DoRemoveProcessor( ctx ); + + /* B */ + _SMP_barrier_Wait( &ctx->barrier, &barrier_state, 2 ); + + /* Clean up all used resources */ + SetSelfPriority( PRIO_NORMAL ); + SendEvents( ctx->worker_id[ WORKER_A ], EVENT_RELEASE ); + T_busy(100000); +} + +static void Worker( rtems_task_argument arg, WorkerIndex worker ) +{ + Context *ctx; + + ctx = (Context *) arg; + + while ( true ) { + rtems_event_set events; + + events = ReceiveAnyEvents(); + + if ( ( events & EVENT_SYNC_RUNNER ) != 0 ) { + SendEvents( ctx->runner_id, EVENT_SYNC_RUNNER ); } - case RtemsSchedulerReqRemoveProcessor_Pre_CPUState_InUse: { - /* - * While the processor associated with the ``cpu_index`` parameter is - * owned by the scheduler specified by the ``scheduler_id`` parameter, - * while the scheduler is used by at least one task as its home scheduler - * and the affinity set of this task requires the processor specified by - * the ``cpu_index`` parameter. - */ - /* Set by prologue */ - break; + if ( ( events & EVENT_OBTAIN ) != 0 ) { + ObtainMutex( ctx->mutex_id ); } - case RtemsSchedulerReqRemoveProcessor_Pre_CPUState_NotOwned: { - /* - * While the processor associated with the ``cpu_index`` parameter is not - * owned by the scheduler specified by the ``scheduler_id`` parameter. - */ - ctx->scheduler_id = ctx->scheduler_a_id; - ctx->cpu_to_remove = 1; - break; + if ( ( events & EVENT_RELEASE ) != 0 ) { + ReleaseMutex( ctx->mutex_id ); } - case RtemsSchedulerReqRemoveProcessor_Pre_CPUState_NA: - break; + if ( ( events & EVENT_STICKY_OBTAIN ) != 0 ) { + ObtainMutex( ctx->sticky_id ); + } + + if ( ( events & EVENT_STICKY_RELEASE ) != 0 ) { + ReleaseMutex( ctx->sticky_id ); + } + + if ( ( events & EVENT_RESTART ) != 0 ) { + rtems_status_code sc; + + T_eq_u32( rtems_scheduler_get_processor(), 0 ); + SetPriority( ctx->runner_id, PRIO_VERY_HIGH ); + T_eq_u32( rtems_scheduler_get_processor(), 1 ); + + if ( !ctx->last ) { + SetScheduler( ctx->worker_id[ WORKER_C ], SCHEDULER_A_ID, PRIO_LOW ); + RemoveProcessor( SCHEDULER_C_ID, 2 ); + AddProcessor( SCHEDULER_B_ID, 2 ); + } + + WrapThreadQueueExtract( + &ctx->wrap_tq_ctx, + GetThread( ctx->worker_id[ WORKER_B ] ) + ); + + sc = rtems_task_restart( + ctx->worker_id[ WORKER_B ], + (rtems_task_argument) ctx + ); + T_rsc_success( sc ); + + T_eq_u32( rtems_scheduler_get_processor(), 0 ); + + if ( !ctx->last ) { + RemoveProcessor( SCHEDULER_B_ID, 2 ); + AddProcessor( SCHEDULER_C_ID, 2 ); + SetScheduler( ctx->worker_id[ WORKER_C ], SCHEDULER_C_ID, PRIO_NORMAL ); + } + } + + if ( ( events & EVENT_BUSY ) != 0 ) { + while ( ctx->busy[ worker ] ) { + ++ctx->busy_counter[ worker ]; + } + } + + if ( ( events & EVENT_SYNC_RUNNER_LATE ) != 0 ) { + SendEvents( ctx->runner_id, EVENT_SYNC_RUNNER ); + } } } +static void WorkerA( rtems_task_argument arg ) +{ + Worker( arg, WORKER_A ); +} + +static void WorkerB( rtems_task_argument arg ) +{ + Worker( arg, WORKER_B ); +} + +static void WorkerC( rtems_task_argument arg ) +{ + Worker( arg, WORKER_C ); +} +#endif + static void RtemsSchedulerReqRemoveProcessor_Pre_Id_Prepare( RtemsSchedulerReqRemoveProcessor_Context *ctx, RtemsSchedulerReqRemoveProcessor_Pre_Id state @@ -289,7 +608,7 @@ static void RtemsSchedulerReqRemoveProcessor_Pre_Id_Prepare( /* * While the ``scheduler_id`` parameter is associated with a scheduler. */ - ctx->id = ctx->scheduler_id; + ctx->id = SCHEDULER_A_ID; break; } @@ -309,7 +628,7 @@ static void RtemsSchedulerReqRemoveProcessor_Pre_CPUIndex_Prepare( * While the ``cpu_index`` parameter is less than the configured * processor maximum. */ - ctx->cpu_index = ctx->cpu_to_remove; + ctx->cpu_index = 0; break; } @@ -327,6 +646,190 @@ static void RtemsSchedulerReqRemoveProcessor_Pre_CPUIndex_Prepare( } } +static void RtemsSchedulerReqRemoveProcessor_Pre_Owned_Prepare( + RtemsSchedulerReqRemoveProcessor_Context *ctx, + RtemsSchedulerReqRemoveProcessor_Pre_Owned state +) +{ + switch ( state ) { + case RtemsSchedulerReqRemoveProcessor_Pre_Owned_Yes: { + /* + * While the processor specified by the ``cpu_index`` parameter is owned + * by the scheduler specified by the ``scheduler_id`` parameter. + */ + ctx->owned = true; + break; + } + + case RtemsSchedulerReqRemoveProcessor_Pre_Owned_No: { + /* + * While the processor specified by the ``cpu_index`` parameter is not + * owned by the scheduler specified by the ``scheduler_id`` parameter. + */ + ctx->owned = false; + break; + } + + case RtemsSchedulerReqRemoveProcessor_Pre_Owned_NA: + break; + } +} + +static void RtemsSchedulerReqRemoveProcessor_Pre_Last_Prepare( + RtemsSchedulerReqRemoveProcessor_Context *ctx, + RtemsSchedulerReqRemoveProcessor_Pre_Last state +) +{ + switch ( state ) { + case RtemsSchedulerReqRemoveProcessor_Pre_Last_Yes: { + /* + * While the processor specified by the ``cpu_index`` parameter is the + * last processor owned by the scheduler specified by the + * ``scheduler_id`` parameter. + */ + ctx->last = true; + break; + } + + case RtemsSchedulerReqRemoveProcessor_Pre_Last_No: { + /* + * While the processor specified by the ``cpu_index`` parameter is not + * the last processor owned by the scheduler specified by the + * ``scheduler_id`` parameter. + */ + ctx->last = false; + break; + } + + case RtemsSchedulerReqRemoveProcessor_Pre_Last_NA: + break; + } +} + +static void RtemsSchedulerReqRemoveProcessor_Pre_Home_Prepare( + RtemsSchedulerReqRemoveProcessor_Context *ctx, + RtemsSchedulerReqRemoveProcessor_Pre_Home state +) +{ + switch ( state ) { + case RtemsSchedulerReqRemoveProcessor_Pre_Home_Yes: { + /* + * While at least one non-idle task exists which uses the scheduler + * specified by the ``scheduler_id`` parameter as its home scheduler. + */ + ctx->home = true; + break; + } + + case RtemsSchedulerReqRemoveProcessor_Pre_Home_No: { + /* + * While no non-idle task exists which uses the scheduler specified by + * the ``scheduler_id`` parameter as its home scheduler. + */ + ctx->home = false; + break; + } + + case RtemsSchedulerReqRemoveProcessor_Pre_Home_NA: + break; + } +} + +static void RtemsSchedulerReqRemoveProcessor_Pre_RequiredByAffinity_Prepare( + RtemsSchedulerReqRemoveProcessor_Context *ctx, + RtemsSchedulerReqRemoveProcessor_Pre_RequiredByAffinity state +) +{ + switch ( state ) { + case RtemsSchedulerReqRemoveProcessor_Pre_RequiredByAffinity_Yes: { + /* + * While at least one non-idle task which uses the scheduler specified by + * the ``scheduler_id`` parameter as its home scheduler exists those + * processor affinity set requires the processor specified by the + * ``cpu_index`` parameter. + */ + ctx->required_by_affinity = true; + break; + } + + case RtemsSchedulerReqRemoveProcessor_Pre_RequiredByAffinity_No: { + /* + * While no non-idle task which uses the scheduler specified by the + * ``scheduler_id`` parameter as its home scheduler exists those + * processor affinity set requires the processor specified by the + * ``cpu_index`` parameter. + */ + ctx->required_by_affinity = false; + break; + } + + case RtemsSchedulerReqRemoveProcessor_Pre_RequiredByAffinity_NA: + break; + } +} + +static void RtemsSchedulerReqRemoveProcessor_Pre_UsedBy_Prepare( + RtemsSchedulerReqRemoveProcessor_Context *ctx, + RtemsSchedulerReqRemoveProcessor_Pre_UsedBy state +) +{ + switch ( state ) { + case RtemsSchedulerReqRemoveProcessor_Pre_UsedBy_Idle: { + /* + * While the processor specified by the ``cpu_index`` parameter is used + * by an idle task. + */ + ctx->idle = true; + ctx->task = false; + ctx->helping = false; + break; + } + + case RtemsSchedulerReqRemoveProcessor_Pre_UsedBy_Task: { + /* + * While the processor specified by the ``cpu_index`` parameter is used + * by a task task which uses the scheduler specified by the + * ``scheduler_id`` parameter as its home scheduler. + */ + ctx->idle = false; + ctx->task = true; + ctx->helping = false; + break; + } + + case RtemsSchedulerReqRemoveProcessor_Pre_UsedBy_TaskIdle: { + /* + * While the processor specified by the ``cpu_index`` parameter is used + * by an idle task on behalf of a task task which uses the scheduler + * specified by the ``scheduler_id`` parameter as its home scheduler. + */ + ctx->idle = true; + ctx->task = true; + ctx->helping = false; + break; + } + + case RtemsSchedulerReqRemoveProcessor_Pre_UsedBy_Helping: { + /* + * While the processor specified by the ``cpu_index`` parameter is used + * by a task task which uses the scheduler specified by the + * ``scheduler_id`` parameter as a helping scheduler. + */ + if ( !ctx->last && rtems_scheduler_get_processor_maximum() < 3 ) { + ctx->Map.skip = true; + } else { + ctx->idle = false; + ctx->task = false; + ctx->helping = true; + } + break; + } + + case RtemsSchedulerReqRemoveProcessor_Pre_UsedBy_NA: + break; + } +} + static void RtemsSchedulerReqRemoveProcessor_Post_Status_Check( RtemsSchedulerReqRemoveProcessor_Context *ctx, RtemsSchedulerReqRemoveProcessor_Post_Status state @@ -386,11 +889,27 @@ static void RtemsSchedulerReqRemoveProcessor_Post_Removed_Check( * removed from the scheduler specified by the ``scheduler_id`` by the * rtems_scheduler_remove_processor() call. */ - T_eq_sz( ctx->scheduler_log.header.recorded, 1 ); - T_eq_int( - ctx->scheduler_log.events[ 0 ].operation, - T_SCHEDULER_REMOVE_PROCESSOR - ); + if ( ctx->home && ctx->helping ) { + T_eq_sz( ctx->scheduler_log.header.recorded, 3 ); + T_eq_int( + ctx->scheduler_log.events[ 0 ].operation, + T_SCHEDULER_REMOVE_PROCESSOR + ); + T_eq_int( + ctx->scheduler_log.events[ 1 ].operation, + T_SCHEDULER_ASK_FOR_HELP + ); + T_eq_int( + ctx->scheduler_log.events[ 2 ].operation, + T_SCHEDULER_ASK_FOR_HELP + ); + } else { + T_eq_sz( ctx->scheduler_log.header.recorded, 1 ); + T_eq_int( + ctx->scheduler_log.events[ 0 ].operation, + T_SCHEDULER_REMOVE_PROCESSOR + ); + } break; } @@ -412,19 +931,59 @@ static void RtemsSchedulerReqRemoveProcessor_Setup( RtemsSchedulerReqRemoveProcessor_Context *ctx ) { - rtems_status_code sc; - - sc = rtems_scheduler_ident( - TEST_SCHEDULER_A_NAME, - &ctx->scheduler_a_id + #if defined(RTEMS_SMP) + rtems_status_code sc; + rtems_task_priority priority; + + ctx->runner_id = rtems_task_self(); + ctx->mutex_id = CreateMutex(); + + sc = rtems_semaphore_create( + rtems_build_name( 'S', 'T', 'K', 'Y' ), + 1, + RTEMS_BINARY_SEMAPHORE | RTEMS_PRIORITY | + RTEMS_MULTIPROCESSOR_RESOURCE_SHARING, + PRIO_NORMAL, + &ctx->sticky_id ); T_rsc_success( sc ); - #if defined(RTEMS_SMP) - sc = rtems_scheduler_ident( TEST_SCHEDULER_B_NAME, &ctx->scheduler_b_id ); + sc = rtems_semaphore_set_priority( + ctx->sticky_id, + SCHEDULER_B_ID, + PRIO_NORMAL, + &priority + ); T_rsc_success( sc ); - #else - ctx->scheduler_b_id = INVALID_ID; + + if ( rtems_scheduler_get_processor_maximum() >= 3 ) { + sc = rtems_semaphore_set_priority( + ctx->sticky_id, + SCHEDULER_C_ID, + PRIO_LOW, + &priority + ); + T_rsc_success( sc ); + + ctx->worker_id[ WORKER_C ] = CreateTask( "WRKC", PRIO_NORMAL ); + SetScheduler( ctx->worker_id[ WORKER_C ], SCHEDULER_C_ID, PRIO_NORMAL ); + StartTask( ctx->worker_id[ WORKER_C ], WorkerC, ctx ); + + if ( rtems_scheduler_get_processor_maximum() >= 4 ) { + RemoveProcessor( SCHEDULER_C_ID, 3 ); + } + } + + SetSelfPriority( PRIO_NORMAL ); + SetSelfAffinityOne( 0 ); + + ctx->worker_id[ WORKER_A ] = CreateTask( "WRKA", PRIO_HIGH ); + StartTask( ctx->worker_id[ WORKER_A ], WorkerA, ctx ); + + ctx->worker_id[ WORKER_B ] = CreateTask( "WRKB", PRIO_HIGH ); + StartTask( ctx->worker_id[ WORKER_B ], WorkerB, ctx ); + + WrapThreadQueueInitialize( &ctx->wrap_tq_ctx, RequestISR, ctx ); #endif } @@ -437,65 +996,240 @@ static void RtemsSchedulerReqRemoveProcessor_Setup_Wrap( void *arg ) RtemsSchedulerReqRemoveProcessor_Setup( ctx ); } -static void RtemsSchedulerReqRemoveProcessor_Action( +static void RtemsSchedulerReqRemoveProcessor_Teardown( RtemsSchedulerReqRemoveProcessor_Context *ctx ) { - T_scheduler_log *log; + #if defined(RTEMS_SMP) + DeleteTask( ctx->worker_id[ WORKER_A ] ); + DeleteTask( ctx->worker_id[ WORKER_B ] ); + DeleteTask( ctx->worker_id[ WORKER_C ] ); + DeleteMutex( ctx->mutex_id ); + DeleteMutex( ctx->sticky_id ); + WrapThreadQueueDestroy( &ctx->wrap_tq_ctx ); + + if ( rtems_scheduler_get_processor_maximum() >= 4 ) { + AddProcessor( SCHEDULER_C_ID, 3 ); + } - log = T_scheduler_record_2( &ctx->scheduler_log ); - T_null( log ); + RestoreRunnerPriority(); + SetSelfAffinityAll(); + #endif +} - ctx->status = rtems_scheduler_remove_processor( ctx->id, ctx->cpu_index ); +static void RtemsSchedulerReqRemoveProcessor_Teardown_Wrap( void *arg ) +{ + RtemsSchedulerReqRemoveProcessor_Context *ctx; - log = T_scheduler_record( NULL ); - T_eq_ptr( &log->header, &ctx->scheduler_log.header ); + ctx = arg; + ctx->Map.in_action_loop = false; + RtemsSchedulerReqRemoveProcessor_Teardown( ctx ); } -static void RtemsSchedulerReqRemoveProcessor_Cleanup( +static void RtemsSchedulerReqRemoveProcessor_Prepare( RtemsSchedulerReqRemoveProcessor_Context *ctx ) { + ctx->status = RTEMS_NOT_IMPLEMENTED; +} + +static void RtemsSchedulerReqRemoveProcessor_Action( + RtemsSchedulerReqRemoveProcessor_Context *ctx +) +{ + if ( + ctx->id == INVALID_ID || + ctx->cpu_index == rtems_configuration_get_maximum_processors() || + ( ctx->owned && ctx->last && ctx->home && ctx->required_by_affinity && + ( ctx->task || ctx->idle ) ) + ) { + DoRemoveProcessor( ctx ); + } else { #if defined(RTEMS_SMP) - if ( ctx->status == RTEMS_SUCCESSFUL ) { - rtems_status_code sc; + if ( ctx->owned && !ctx->home && ctx->helping ) { + RemoveWithHelpingOnly( ctx ); + } else { + if ( ctx->owned ) { + rtems_id worker_a; + rtems_id worker_b; + + worker_a = ctx->worker_id[ WORKER_A ]; + worker_b = ctx->worker_id[ WORKER_B ]; + + ctx->cpu_index = 1; + + if ( ctx->last ) { + ctx->id = SCHEDULER_B_ID; + } else { + RemoveProcessor( SCHEDULER_B_ID, 1 ); + AddProcessor( SCHEDULER_A_ID, 1 ); + } - sc = rtems_scheduler_add_processor( - ctx->scheduler_id, - ctx->cpu_to_remove - ); - T_rsc_success( sc ); - } + if ( ctx->home ) { + SetScheduler( worker_a, ctx->id, PRIO_LOW ); + + if ( ctx->required_by_affinity ) { + SetAffinityOne( worker_a, 1 ); + } else { + SetAffinityAll( worker_a ); + } + } + + if ( ctx->idle ) { + if ( ctx->task ) { + SendAndSync( ctx, WORKER_A, EVENT_STICKY_OBTAIN ); + SuspendTask( worker_a ); + } + } else if ( ctx->task ) { + MakeBusy( ctx, WORKER_A ); + } else if ( ctx->helping ) { + T_true( ctx->home ); + + if ( ctx->last ) { + SendEvents( worker_b, EVENT_OBTAIN ); + SetPriority( worker_b, PRIO_LOW ); + } else { + SetScheduler( worker_b, SCHEDULER_C_ID, PRIO_LOW ); + SendAndSync( ctx, WORKER_B, EVENT_OBTAIN ); + MakeBusyAndSync( ctx, WORKER_C ); + } + + SendAndSync( ctx, WORKER_A, EVENT_OBTAIN ); + MakeBusy( ctx, WORKER_B ); + WaitForBusy( ctx, WORKER_B ); + } + + DoRemoveProcessor( ctx ); + + if ( ctx->idle ) { + if ( ctx->task ) { + ResumeTask( worker_a ); + SendAndSync( ctx, WORKER_A, EVENT_STICKY_RELEASE ); + } + } else if ( ctx->task ) { + StopBusyAndWait( ctx, WORKER_A ); + } else if ( ctx->helping ) { + StopBusy( ctx, WORKER_B ); + + if ( ctx->last ) { + SetPriority( worker_b, PRIO_HIGH ); + SendEvents( worker_b, EVENT_RELEASE ); + } else { + StopBusyAndWait( ctx, WORKER_C ); + SendAndSync( ctx, WORKER_B, EVENT_RELEASE ); + SetScheduler( worker_b, SCHEDULER_A_ID, PRIO_HIGH ); + } + + WaitForExecutionStop( worker_b ); + SendAndSync( ctx, WORKER_A, EVENT_RELEASE ); + } + + SetAffinityAll( worker_a ); + SetScheduler( worker_a, SCHEDULER_A_ID, PRIO_HIGH ); + + if ( !ctx->last ) { + RemoveProcessor( SCHEDULER_A_ID, 1 ); + AddProcessor( SCHEDULER_B_ID, 1 ); + } + } else { + ctx->id = SCHEDULER_B_ID; + DoRemoveProcessor( ctx ); + } + } + #else + T_unreachable(); #endif + } } static const RtemsSchedulerReqRemoveProcessor_Entry RtemsSchedulerReqRemoveProcessor_Entries[] = { - { 0, 1, 0, 0, RtemsSchedulerReqRemoveProcessor_Post_Status_InvId, +#if !defined(RTEMS_SMP) + { 1, 0, 0, 0, 0, 0, 0, 0, RtemsSchedulerReqRemoveProcessor_Post_Status_NA, + RtemsSchedulerReqRemoveProcessor_Post_Removed_NA }, +#else + { 0, 0, 0, 1, 1, 1, 1, 1, RtemsSchedulerReqRemoveProcessor_Post_Status_InvId, RtemsSchedulerReqRemoveProcessor_Post_Removed_Nop }, - { 0, 1, 0, 0, RtemsSchedulerReqRemoveProcessor_Post_Status_InvNum, +#endif +#if !defined(RTEMS_SMP) + { 1, 0, 0, 0, 0, 0, 0, 0, RtemsSchedulerReqRemoveProcessor_Post_Status_NA, + RtemsSchedulerReqRemoveProcessor_Post_Removed_NA }, +#else + { 0, 0, 0, 1, 1, 1, 1, 1, + RtemsSchedulerReqRemoveProcessor_Post_Status_InvNum, RtemsSchedulerReqRemoveProcessor_Post_Removed_Nop }, -#if defined(RTEMS_SMP) - { 0, 0, 0, 0, RtemsSchedulerReqRemoveProcessor_Post_Status_Ok, +#endif +#if !defined(RTEMS_SMP) + { 1, 0, 0, 0, 0, 0, 0, 0, RtemsSchedulerReqRemoveProcessor_Post_Status_NA, + RtemsSchedulerReqRemoveProcessor_Post_Removed_NA }, +#else + { 0, 0, 0, 0, 1, 1, 1, 1, + RtemsSchedulerReqRemoveProcessor_Post_Status_InvNum, + RtemsSchedulerReqRemoveProcessor_Post_Removed_Nop }, +#endif +#if !defined(RTEMS_SMP) + { 1, 0, 0, 0, 0, 0, 0, 0, RtemsSchedulerReqRemoveProcessor_Post_Status_NA, + RtemsSchedulerReqRemoveProcessor_Post_Removed_NA }, +#else + { 1, 0, 0, 0, 0, 0, 1, 0, RtemsSchedulerReqRemoveProcessor_Post_Status_NA, + RtemsSchedulerReqRemoveProcessor_Post_Removed_NA }, +#endif +#if !defined(RTEMS_SMP) + { 1, 0, 0, 0, 0, 0, 0, 0, RtemsSchedulerReqRemoveProcessor_Post_Status_NA, + RtemsSchedulerReqRemoveProcessor_Post_Removed_NA }, +#else + { 0, 0, 0, 0, 0, 0, 0, 0, RtemsSchedulerReqRemoveProcessor_Post_Status_InUse, + RtemsSchedulerReqRemoveProcessor_Post_Removed_Nop }, +#endif +#if !defined(RTEMS_SMP) + { 1, 0, 0, 0, 0, 0, 0, 0, RtemsSchedulerReqRemoveProcessor_Post_Status_NA, + RtemsSchedulerReqRemoveProcessor_Post_Removed_NA }, +#else + { 0, 0, 0, 0, 0, 0, 1, 0, RtemsSchedulerReqRemoveProcessor_Post_Status_Ok, RtemsSchedulerReqRemoveProcessor_Post_Removed_Yes }, +#endif + { 0, 0, 0, 1, 1, 1, 1, 1, RtemsSchedulerReqRemoveProcessor_Post_Status_InvId, + RtemsSchedulerReqRemoveProcessor_Post_Removed_Nop }, +#if !defined(RTEMS_SMP) + { 1, 0, 0, 0, 0, 0, 0, 0, RtemsSchedulerReqRemoveProcessor_Post_Status_NA, + RtemsSchedulerReqRemoveProcessor_Post_Removed_NA }, #else - { 1, 0, 0, 0, RtemsSchedulerReqRemoveProcessor_Post_Status_NA, + { 1, 0, 0, 0, 0, 0, 0, 0, RtemsSchedulerReqRemoveProcessor_Post_Status_NA, RtemsSchedulerReqRemoveProcessor_Post_Removed_NA }, #endif - { 0, 0, 0, 0, RtemsSchedulerReqRemoveProcessor_Post_Status_InUse, +#if !defined(RTEMS_SMP) + { 1, 0, 0, 0, 0, 0, 0, 0, RtemsSchedulerReqRemoveProcessor_Post_Status_NA, + RtemsSchedulerReqRemoveProcessor_Post_Removed_NA }, +#else + { 0, 0, 0, 0, 0, 0, 0, 0, RtemsSchedulerReqRemoveProcessor_Post_Status_Ok, + RtemsSchedulerReqRemoveProcessor_Post_Removed_Yes }, +#endif + { 0, 0, 0, 0, 0, 0, 0, 0, RtemsSchedulerReqRemoveProcessor_Post_Status_InUse, RtemsSchedulerReqRemoveProcessor_Post_Removed_Nop }, -#if defined(RTEMS_SMP) - { 0, 0, 0, 0, RtemsSchedulerReqRemoveProcessor_Post_Status_InvNum, - RtemsSchedulerReqRemoveProcessor_Post_Removed_Nop } +#if !defined(RTEMS_SMP) + { 1, 0, 0, 0, 0, 0, 0, 0, RtemsSchedulerReqRemoveProcessor_Post_Status_NA, + RtemsSchedulerReqRemoveProcessor_Post_Removed_NA }, #else - { 1, 0, 0, 0, RtemsSchedulerReqRemoveProcessor_Post_Status_NA, - RtemsSchedulerReqRemoveProcessor_Post_Removed_NA } + { 0, 0, 0, 0, 0, 0, 1, 0, RtemsSchedulerReqRemoveProcessor_Post_Status_InUse, + RtemsSchedulerReqRemoveProcessor_Post_Removed_Nop }, #endif + { 0, 0, 0, 1, 1, 1, 1, 1, + RtemsSchedulerReqRemoveProcessor_Post_Status_InvNum, + RtemsSchedulerReqRemoveProcessor_Post_Removed_Nop } }; static const uint8_t RtemsSchedulerReqRemoveProcessor_Map[] = { - 0, 0, 2, 1, 0, 0, 3, 1, 0, 0, 4, 1 + 6, 6, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, + 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, + 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 6, 6, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, + 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, + 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 9, 9, + 4, 4, 7, 7, 7, 7, 5, 3, 3, 10, 5, 3, 3, 10, 4, 4, 4, 4, 8, 8, 8, 8, 5, 3, 3, + 5, 5, 3, 3, 5, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, + 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 11, 11, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, + 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, + 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1 }; static size_t RtemsSchedulerReqRemoveProcessor_Scope( @@ -523,11 +1257,42 @@ static size_t RtemsSchedulerReqRemoveProcessor_Scope( static T_fixture RtemsSchedulerReqRemoveProcessor_Fixture = { .setup = RtemsSchedulerReqRemoveProcessor_Setup_Wrap, .stop = NULL, - .teardown = NULL, + .teardown = RtemsSchedulerReqRemoveProcessor_Teardown_Wrap, .scope = RtemsSchedulerReqRemoveProcessor_Scope, .initial_context = &RtemsSchedulerReqRemoveProcessor_Instance }; +static const uint8_t RtemsSchedulerReqRemoveProcessor_Weights[] = { + 128, 64, 32, 16, 8, 4, 1 +}; + +static void RtemsSchedulerReqRemoveProcessor_Skip( + RtemsSchedulerReqRemoveProcessor_Context *ctx, + size_t index +) +{ + switch ( index + 1 ) { + case 1: + ctx->Map.pci[ 1 ] = RtemsSchedulerReqRemoveProcessor_Pre_CPUIndex_NA - 1; + /* Fall through */ + case 2: + ctx->Map.pci[ 2 ] = RtemsSchedulerReqRemoveProcessor_Pre_Owned_NA - 1; + /* Fall through */ + case 3: + ctx->Map.pci[ 3 ] = RtemsSchedulerReqRemoveProcessor_Pre_Last_NA - 1; + /* Fall through */ + case 4: + ctx->Map.pci[ 4 ] = RtemsSchedulerReqRemoveProcessor_Pre_Home_NA - 1; + /* Fall through */ + case 5: + ctx->Map.pci[ 5 ] = RtemsSchedulerReqRemoveProcessor_Pre_RequiredByAffinity_NA - 1; + /* Fall through */ + case 6: + ctx->Map.pci[ 6 ] = RtemsSchedulerReqRemoveProcessor_Pre_UsedBy_NA - 1; + break; + } +} + static inline RtemsSchedulerReqRemoveProcessor_Entry RtemsSchedulerReqRemoveProcessor_PopEntry( RtemsSchedulerReqRemoveProcessor_Context *ctx @@ -535,8 +1300,21 @@ RtemsSchedulerReqRemoveProcessor_PopEntry( { size_t index; - index = ctx->Map.index; + if ( ctx->Map.skip ) { + size_t i; + + ctx->Map.skip = false; + index = 0; + + for ( i = 0; i < 7; ++i ) { + index += RtemsSchedulerReqRemoveProcessor_Weights[ i ] * ctx->Map.pci[ i ]; + } + } else { + index = ctx->Map.index; + } + ctx->Map.index = index + 1; + return RtemsSchedulerReqRemoveProcessor_Entries[ RtemsSchedulerReqRemoveProcessor_Map[ index ] ]; @@ -546,29 +1324,66 @@ static void RtemsSchedulerReqRemoveProcessor_SetPreConditionStates( RtemsSchedulerReqRemoveProcessor_Context *ctx ) { - if ( ctx->Map.entry.Pre_CPUState_NA ) { - ctx->Map.pcs[ 0 ] = RtemsSchedulerReqRemoveProcessor_Pre_CPUState_NA; + ctx->Map.pcs[ 0 ] = ctx->Map.pci[ 0 ]; + ctx->Map.pcs[ 1 ] = ctx->Map.pci[ 1 ]; + + if ( ctx->Map.entry.Pre_Owned_NA ) { + ctx->Map.pcs[ 2 ] = RtemsSchedulerReqRemoveProcessor_Pre_Owned_NA; } else { - ctx->Map.pcs[ 0 ] = ctx->Map.pci[ 0 ]; + ctx->Map.pcs[ 2 ] = ctx->Map.pci[ 2 ]; } - ctx->Map.pcs[ 1 ] = ctx->Map.pci[ 1 ]; - ctx->Map.pcs[ 2 ] = ctx->Map.pci[ 2 ]; + if ( ctx->Map.entry.Pre_Last_NA ) { + ctx->Map.pcs[ 3 ] = RtemsSchedulerReqRemoveProcessor_Pre_Last_NA; + } else { + ctx->Map.pcs[ 3 ] = ctx->Map.pci[ 3 ]; + } + + if ( ctx->Map.entry.Pre_Home_NA ) { + ctx->Map.pcs[ 4 ] = RtemsSchedulerReqRemoveProcessor_Pre_Home_NA; + } else { + ctx->Map.pcs[ 4 ] = ctx->Map.pci[ 4 ]; + } + + if ( ctx->Map.entry.Pre_RequiredByAffinity_NA ) { + ctx->Map.pcs[ 5 ] = RtemsSchedulerReqRemoveProcessor_Pre_RequiredByAffinity_NA; + } else { + ctx->Map.pcs[ 5 ] = ctx->Map.pci[ 5 ]; + } + + if ( ctx->Map.entry.Pre_UsedBy_NA ) { + ctx->Map.pcs[ 6 ] = RtemsSchedulerReqRemoveProcessor_Pre_UsedBy_NA; + } else { + ctx->Map.pcs[ 6 ] = ctx->Map.pci[ 6 ]; + } } static void RtemsSchedulerReqRemoveProcessor_TestVariant( RtemsSchedulerReqRemoveProcessor_Context *ctx ) { - RtemsSchedulerReqRemoveProcessor_Pre_CPUState_Prepare( + RtemsSchedulerReqRemoveProcessor_Pre_Id_Prepare( ctx, ctx->Map.pcs[ 0 ] ); + RtemsSchedulerReqRemoveProcessor_Pre_CPUIndex_Prepare( ctx, - ctx->Map.pcs[ 0 ] + ctx->Map.pcs[ 1 ] ); - RtemsSchedulerReqRemoveProcessor_Pre_Id_Prepare( ctx, ctx->Map.pcs[ 1 ] ); - RtemsSchedulerReqRemoveProcessor_Pre_CPUIndex_Prepare( + RtemsSchedulerReqRemoveProcessor_Pre_Owned_Prepare( ctx, ctx->Map.pcs[ 2 ] ); + RtemsSchedulerReqRemoveProcessor_Pre_Last_Prepare( ctx, ctx->Map.pcs[ 3 ] ); + RtemsSchedulerReqRemoveProcessor_Pre_Home_Prepare( ctx, ctx->Map.pcs[ 4 ] ); + RtemsSchedulerReqRemoveProcessor_Pre_RequiredByAffinity_Prepare( + ctx, + ctx->Map.pcs[ 5 ] + ); + RtemsSchedulerReqRemoveProcessor_Pre_UsedBy_Prepare( ctx, - ctx->Map.pcs[ 2 ] + ctx->Map.pcs[ 6 ] ); + + if ( ctx->Map.skip ) { + RtemsSchedulerReqRemoveProcessor_Skip( ctx, 6 ); + return; + } + RtemsSchedulerReqRemoveProcessor_Action( ctx ); RtemsSchedulerReqRemoveProcessor_Post_Status_Check( ctx, @@ -593,31 +1408,58 @@ T_TEST_CASE_FIXTURE( ctx = T_fixture_context(); ctx->Map.in_action_loop = true; ctx->Map.index = 0; + ctx->Map.skip = false; for ( - ctx->Map.pci[ 0 ] = RtemsSchedulerReqRemoveProcessor_Pre_CPUState_Idle; - ctx->Map.pci[ 0 ] < RtemsSchedulerReqRemoveProcessor_Pre_CPUState_NA; + ctx->Map.pci[ 0 ] = RtemsSchedulerReqRemoveProcessor_Pre_Id_Invalid; + ctx->Map.pci[ 0 ] < RtemsSchedulerReqRemoveProcessor_Pre_Id_NA; ++ctx->Map.pci[ 0 ] ) { for ( - ctx->Map.pci[ 1 ] = RtemsSchedulerReqRemoveProcessor_Pre_Id_Invalid; - ctx->Map.pci[ 1 ] < RtemsSchedulerReqRemoveProcessor_Pre_Id_NA; + ctx->Map.pci[ 1 ] = RtemsSchedulerReqRemoveProcessor_Pre_CPUIndex_Valid; + ctx->Map.pci[ 1 ] < RtemsSchedulerReqRemoveProcessor_Pre_CPUIndex_NA; ++ctx->Map.pci[ 1 ] ) { for ( - ctx->Map.pci[ 2 ] = RtemsSchedulerReqRemoveProcessor_Pre_CPUIndex_Valid; - ctx->Map.pci[ 2 ] < RtemsSchedulerReqRemoveProcessor_Pre_CPUIndex_NA; + ctx->Map.pci[ 2 ] = RtemsSchedulerReqRemoveProcessor_Pre_Owned_Yes; + ctx->Map.pci[ 2 ] < RtemsSchedulerReqRemoveProcessor_Pre_Owned_NA; ++ctx->Map.pci[ 2 ] ) { - ctx->Map.entry = RtemsSchedulerReqRemoveProcessor_PopEntry( ctx ); - - if ( ctx->Map.entry.Skip ) { - continue; + for ( + ctx->Map.pci[ 3 ] = RtemsSchedulerReqRemoveProcessor_Pre_Last_Yes; + ctx->Map.pci[ 3 ] < RtemsSchedulerReqRemoveProcessor_Pre_Last_NA; + ++ctx->Map.pci[ 3 ] + ) { + for ( + ctx->Map.pci[ 4 ] = RtemsSchedulerReqRemoveProcessor_Pre_Home_Yes; + ctx->Map.pci[ 4 ] < RtemsSchedulerReqRemoveProcessor_Pre_Home_NA; + ++ctx->Map.pci[ 4 ] + ) { + for ( + ctx->Map.pci[ 5 ] = RtemsSchedulerReqRemoveProcessor_Pre_RequiredByAffinity_Yes; + ctx->Map.pci[ 5 ] < RtemsSchedulerReqRemoveProcessor_Pre_RequiredByAffinity_NA; + ++ctx->Map.pci[ 5 ] + ) { + for ( + ctx->Map.pci[ 6 ] = RtemsSchedulerReqRemoveProcessor_Pre_UsedBy_Idle; + ctx->Map.pci[ 6 ] < RtemsSchedulerReqRemoveProcessor_Pre_UsedBy_NA; + ++ctx->Map.pci[ 6 ] + ) { + ctx->Map.entry = RtemsSchedulerReqRemoveProcessor_PopEntry( + ctx + ); + + if ( ctx->Map.entry.Skip ) { + continue; + } + + RtemsSchedulerReqRemoveProcessor_SetPreConditionStates( ctx ); + RtemsSchedulerReqRemoveProcessor_Prepare( ctx ); + RtemsSchedulerReqRemoveProcessor_TestVariant( ctx ); + } + } + } } - - RtemsSchedulerReqRemoveProcessor_SetPreConditionStates( ctx ); - RtemsSchedulerReqRemoveProcessor_TestVariant( ctx ); - RtemsSchedulerReqRemoveProcessor_Cleanup( ctx ); } } } |