/** * @file * * @brief Thread Queue Operations * @ingroup ScoreThreadQ */ /* * COPYRIGHT (c) 1989-2014. * On-Line Applications Research Corporation (OAR). * * Copyright (c) 2015, 2016 embedded brains GmbH. * * The license and distribution terms for this file may be * found in the file LICENSE in this distribution or at * http://www.rtems.org/license/LICENSE. */ #if HAVE_CONFIG_H #include "config.h" #endif #include #include #include #include #include #include #define THREAD_QUEUE_INTEND_TO_BLOCK \ (THREAD_WAIT_CLASS_OBJECT | THREAD_WAIT_STATE_INTEND_TO_BLOCK) #define THREAD_QUEUE_BLOCKED \ (THREAD_WAIT_CLASS_OBJECT | THREAD_WAIT_STATE_BLOCKED) #define THREAD_QUEUE_READY_AGAIN \ (THREAD_WAIT_CLASS_OBJECT | THREAD_WAIT_STATE_READY_AGAIN) #if defined(RTEMS_SMP) /* * A global registry of active thread queue links is used to provide deadlock * detection on SMP configurations. This is simple to implement and no * additional storage is required for the thread queues. The disadvantage is * the global registry is not scalable and may lead to lock contention. * However, the registry is only used in case of nested resource conflicts. In * this case, the application is already in trouble. */ typedef struct { ISR_lock_Control Lock; RBTree_Control Links; } Thread_queue_Links; static Thread_queue_Links _Thread_queue_Links = { ISR_LOCK_INITIALIZER( "Thread Queue Links" ), RBTREE_INITIALIZER_EMPTY( _Thread_queue_Links.Links ) }; static bool _Thread_queue_Link_equal( const void *left, const RBTree_Node *right ) { const Thread_queue_Queue *the_left; const Thread_queue_Link *the_right; the_left = left; the_right = (Thread_queue_Link *) right; return the_left == the_right->source; } static bool _Thread_queue_Link_less( const void *left, const RBTree_Node *right ) { const Thread_queue_Queue *the_left; const Thread_queue_Link *the_right; the_left = left; the_right = (Thread_queue_Link *) right; return (uintptr_t) the_left < (uintptr_t) the_right->source; } static void *_Thread_queue_Link_map( RBTree_Node *node ) { return node; } static Thread_queue_Link *_Thread_queue_Link_find( Thread_queue_Links *links, Thread_queue_Queue *source ) { return _RBTree_Find_inline( &links->Links, source, _Thread_queue_Link_equal, _Thread_queue_Link_less, _Thread_queue_Link_map ); } static bool _Thread_queue_Link_add( Thread_queue_Link *link, Thread_queue_Queue *source, Thread_queue_Queue *target ) { Thread_queue_Links *links; Thread_queue_Queue *recursive_target; ISR_lock_Context lock_context; link->source = source; link->target = target; links = &_Thread_queue_Links; recursive_target = target; _ISR_lock_Acquire( &links->Lock, &lock_context ); while ( true ) { Thread_queue_Link *recursive_link; recursive_link = _Thread_queue_Link_find( links, recursive_target ); if ( recursive_link == NULL ) { break; } recursive_target = recursive_link->target; if ( recursive_target == source ) { _ISR_lock_Release( &links->Lock, &lock_context ); return false; } } _RBTree_Insert_inline( &links->Links, &link->Registry_node, source, _Thread_queue_Link_less ); _ISR_lock_Release( &links->Lock, &lock_context ); return true; } static void _Thread_queue_Link_remove( Thread_queue_Link *link ) { Thread_queue_Links *links; ISR_lock_Context lock_context; links = &_Thread_queue_Links; _ISR_lock_Acquire( &links->Lock, &lock_context ); _RBTree_Extract( &links->Links, &link->Registry_node ); _ISR_lock_Release( &links->Lock, &lock_context ); } #endif #if !defined(RTEMS_SMP) static #endif void _Thread_queue_Path_release_critical( Thread_queue_Context *queue_context ) { #if defined(RTEMS_SMP) Chain_Node *head; Chain_Node *node; head = _Chain_Head( &queue_context->Path.Links ); node = _Chain_Last( &queue_context->Path.Links ); while ( head != node ) { Thread_queue_Link *link; link = THREAD_QUEUE_LINK_OF_PATH_NODE( node ); if ( link->Lock_context.Wait.queue != NULL ) { _Thread_queue_Link_remove( link ); _Thread_Wait_release_queue_critical( link->Lock_context.Wait.queue, &link->Lock_context ); _Thread_Wait_remove_request( link->owner, &link->Lock_context ); } else { _Thread_Wait_release_default_critical( link->owner, &link->Lock_context.Lock_context ); } node = _Chain_Previous( node ); #if defined(RTEMS_DEBUG) _Chain_Set_off_chain( &link->Path_node ); #endif } #else (void) queue_context; #endif } #if defined(RTEMS_SMP) static void _Thread_queue_Path_append_deadlock_thread( Thread_Control *the_thread, Thread_queue_Context *queue_context ) { Thread_Control *deadlock; /* * In case of a deadlock, we must obtain the thread wait default lock for the * first thread on the path that tries to enqueue on a thread queue. This * thread can be identified by the thread wait operations. This lock acquire * is necessary for the timeout and explicit thread priority changes, see * _Thread_Priority_perform_actions(). */ deadlock = NULL; while ( the_thread->Wait.operations != &_Thread_queue_Operations_default ) { the_thread = the_thread->Wait.queue->owner; deadlock = the_thread; } if ( deadlock != NULL ) { Thread_queue_Link *link; link = &queue_context->Path.Deadlock; _Chain_Initialize_node( &link->Path_node ); _Chain_Append_unprotected( &queue_context->Path.Links, &link->Path_node ); link->owner = deadlock; link->Lock_context.Wait.queue = NULL; _Thread_Wait_acquire_default_critical( deadlock, &link->Lock_context.Lock_context ); } } #endif #if !defined(RTEMS_SMP) static #endif bool _Thread_queue_Path_acquire_critical( Thread_queue_Queue *queue, Thread_Control *the_thread, Thread_queue_Context *queue_context ) { Thread_Control *owner; #if defined(RTEMS_SMP) Thread_queue_Link *link; Thread_queue_Queue *target; /* * For an overview please look at the non-SMP part below. We basically do * the same on SMP configurations. The fact that we may have more than one * executing thread and each thread queue has its own SMP lock makes the task * a bit more difficult. We have to avoid deadlocks at SMP lock level, since * this would result in an unrecoverable deadlock of the overall system. */ _Chain_Initialize_empty( &queue_context->Path.Links ); owner = queue->owner; if ( owner == NULL ) { return true; } if ( owner == the_thread ) { return false; } _Chain_Initialize_node( &queue_context->Path.Start.Lock_context.Wait.Gate.Node ); link = &queue_context->Path.Start; _RBTree_Initialize_node( &link->Registry_node ); _Chain_Initialize_node( &link->Path_node ); do { _Chain_Append_unprotected( &queue_context->Path.Links, &link->Path_node ); link->owner = owner; _Thread_Wait_acquire_default_critical( owner, &link->Lock_context.Lock_context ); target = owner->Wait.queue; link->Lock_context.Wait.queue = target; if ( target != NULL ) { if ( _Thread_queue_Link_add( link, queue, target ) ) { _Thread_queue_Gate_add( &owner->Wait.Lock.Pending_requests, &link->Lock_context.Wait.Gate ); _Thread_Wait_release_default_critical( owner, &link->Lock_context.Lock_context ); _Thread_Wait_acquire_queue_critical( target, &link->Lock_context ); if ( link->Lock_context.Wait.queue == NULL ) { _Thread_queue_Link_remove( link ); _Thread_Wait_release_queue_critical( target, &link->Lock_context ); _Thread_Wait_acquire_default_critical( owner, &link->Lock_context.Lock_context ); _Thread_Wait_remove_request_locked( owner, &link->Lock_context ); _Assert( owner->Wait.queue == NULL ); return true; } } else { link->Lock_context.Wait.queue = NULL; _Thread_queue_Path_append_deadlock_thread( owner, queue_context ); return false; } } else { return true; } link = &owner->Wait.Link; queue = target; owner = queue->owner; } while ( owner != NULL ); #else do { owner = queue->owner; if ( owner == NULL ) { return true; } if ( owner == the_thread ) { return false; } queue = owner->Wait.queue; } while ( queue != NULL ); #endif return true; } void _Thread_queue_Enqueue_do_nothing_extra( Thread_queue_Queue *queue, Thread_Control *the_thread, Per_CPU_Control *cpu_self, Thread_queue_Context *queue_context ) { /* Do nothing */ } void _Thread_queue_Deadlock_status( Thread_Control *the_thread ) { the_thread->Wait.return_code = STATUS_DEADLOCK; } void _Thread_queue_Deadlock_fatal( Thread_Control *the_thread ) { _Internal_error( INTERNAL_ERROR_THREAD_QUEUE_DEADLOCK ); } void _Thread_queue_Enqueue( Thread_queue_Queue *queue, const Thread_queue_Operations *operations, Thread_Control *the_thread, Thread_queue_Context *queue_context ) { Per_CPU_Control *cpu_self; bool success; _Assert( queue_context->enqueue_callout != NULL ); #if defined(RTEMS_MULTIPROCESSING) if ( _Thread_MP_Is_receive( the_thread ) && the_thread->receive_packet ) { the_thread = _Thread_MP_Allocate_proxy( queue_context->thread_state ); } #endif _Thread_Wait_claim( the_thread, queue ); if ( !_Thread_queue_Path_acquire_critical( queue, the_thread, queue_context ) ) { _Thread_queue_Path_release_critical( queue_context ); _Thread_Wait_restore_default( the_thread ); _Thread_queue_Queue_release( queue, &queue_context->Lock_context.Lock_context ); _Thread_Wait_tranquilize( the_thread ); _Assert( queue_context->deadlock_callout != NULL ); ( *queue_context->deadlock_callout )( the_thread ); return; } _Thread_queue_Context_clear_priority_updates( queue_context ); _Thread_Wait_claim_finalize( the_thread, operations ); ( *operations->enqueue )( queue, the_thread, queue_context ); _Thread_queue_Path_release_critical( queue_context ); the_thread->Wait.return_code = STATUS_SUCCESSFUL; _Thread_Wait_flags_set( the_thread, THREAD_QUEUE_INTEND_TO_BLOCK ); cpu_self = _Thread_queue_Dispatch_disable( queue_context ); _Thread_queue_Queue_release( queue, &queue_context->Lock_context.Lock_context ); ( *queue_context->enqueue_callout )( queue, the_thread, cpu_self, queue_context ); /* * Set the blocking state for this thread queue in the thread. */ _Thread_Set_state( the_thread, queue_context->thread_state ); /* * At this point thread dispatching is disabled, however, we already released * the thread queue lock. Thus, interrupts or threads on other processors * may already changed our state with respect to the thread queue object. * The request could be satisfied or timed out. This situation is indicated * by the thread wait flags. Other parties must not modify our thread state * as long as we are in the THREAD_QUEUE_INTEND_TO_BLOCK thread wait state, * thus we have to cancel the blocking operation ourself if necessary. */ success = _Thread_Wait_flags_try_change_acquire( the_thread, THREAD_QUEUE_INTEND_TO_BLOCK, THREAD_QUEUE_BLOCKED ); if ( !success ) { _Thread_Remove_timer_and_unblock( the_thread, queue ); } _Thread_Priority_update( queue_context ); _Thread_Dispatch_direct( cpu_self ); } #if defined(RTEMS_SMP) Status_Control _Thread_queue_Enqueue_sticky( Thread_queue_Queue *queue, const Thread_queue_Operations *operations, Thread_Control *the_thread, Thread_queue_Context *queue_context ) { Per_CPU_Control *cpu_self; _Assert( queue_context->enqueue_callout != NULL ); _Thread_Wait_claim( the_thread, queue ); if ( !_Thread_queue_Path_acquire_critical( queue, the_thread, queue_context ) ) { _Thread_queue_Path_release_critical( queue_context ); _Thread_Wait_restore_default( the_thread ); _Thread_queue_Queue_release( queue, &queue_context->Lock_context.Lock_context ); _Thread_Wait_tranquilize( the_thread ); ( *queue_context->deadlock_callout )( the_thread ); return _Thread_Wait_get_status( the_thread ); } _Thread_queue_Context_clear_priority_updates( queue_context ); _Thread_Wait_claim_finalize( the_thread, operations ); ( *operations->enqueue )( queue, the_thread, queue_context ); _Thread_queue_Path_release_critical( queue_context ); the_thread->Wait.return_code = STATUS_SUCCESSFUL; _Thread_Wait_flags_set( the_thread, THREAD_QUEUE_INTEND_TO_BLOCK ); cpu_self = _Thread_queue_Dispatch_disable( queue_context ); _Thread_queue_Queue_release( queue, &queue_context->Lock_context.Lock_context ); if ( cpu_self->thread_dispatch_disable_level != 1 ) { _Internal_error( INTERNAL_ERROR_THREAD_QUEUE_ENQUEUE_STICKY_FROM_BAD_STATE ); } ( *queue_context->enqueue_callout )( queue, the_thread, cpu_self, queue_context ); _Thread_Priority_update( queue_context ); _Thread_Priority_and_sticky_update( the_thread, 1 ); _Thread_Dispatch_enable( cpu_self ); while ( _Thread_Wait_flags_get_acquire( the_thread ) == THREAD_QUEUE_INTEND_TO_BLOCK ) { /* Wait */ } _Thread_Wait_tranquilize( the_thread ); _Thread_Timer_remove( the_thread ); return _Thread_Wait_get_status( the_thread ); } #endif #if defined(RTEMS_MULTIPROCESSING) static bool _Thread_queue_MP_set_callout( Thread_Control *the_thread, const Thread_queue_Context *queue_context ) { Thread_Proxy_control *the_proxy; Thread_queue_MP_callout mp_callout; if ( _Objects_Is_local_id( the_thread->Object.id ) ) { return false; } the_proxy = (Thread_Proxy_control *) the_thread; mp_callout = queue_context->mp_callout; _Assert( mp_callout != NULL ); the_proxy->thread_queue_callout = queue_context->mp_callout; return true; } #endif static bool _Thread_queue_Make_ready_again( Thread_Control *the_thread ) { bool success; bool unblock; /* * We must update the wait flags under protection of the current thread lock, * otherwise a _Thread_Timeout() running on another processor may interfere. */ success = _Thread_Wait_flags_try_change_release( the_thread, THREAD_QUEUE_INTEND_TO_BLOCK, THREAD_QUEUE_READY_AGAIN ); if ( success ) { unblock = false; } else { _Assert( _Thread_Wait_flags_get( the_thread ) == THREAD_QUEUE_BLOCKED ); _Thread_Wait_flags_set( the_thread, THREAD_QUEUE_READY_AGAIN ); unblock = true; } _Thread_Wait_restore_default( the_thread ); return unblock; } bool _Thread_queue_Extract_locked( Thread_queue_Queue *queue, const Thread_queue_Operations *operations, Thread_Control *the_thread, Thread_queue_Context *queue_context ) { #if defined(RTEMS_MULTIPROCESSING) _Thread_queue_MP_set_callout( the_thread, queue_context ); #endif ( *operations->extract )( queue, the_thread, queue_context ); return _Thread_queue_Make_ready_again( the_thread ); } void _Thread_queue_Unblock_critical( bool unblock, Thread_queue_Queue *queue, Thread_Control *the_thread, ISR_lock_Context *lock_context ) { if ( unblock ) { Per_CPU_Control *cpu_self; cpu_self = _Thread_Dispatch_disable_critical( lock_context ); _Thread_queue_Queue_release( queue, lock_context ); _Thread_Remove_timer_and_unblock( the_thread, queue ); _Thread_Dispatch_enable( cpu_self ); } else { _Thread_queue_Queue_release( queue, lock_context ); } } void _Thread_queue_Extract_critical( Thread_queue_Queue *queue, const Thread_queue_Operations *operations, Thread_Control *the_thread, Thread_queue_Context *queue_context ) { bool unblock; unblock = _Thread_queue_Extract_locked( queue, operations, the_thread, queue_context ); _Thread_queue_Unblock_critical( unblock, queue, the_thread, &queue_context->Lock_context.Lock_context ); } void _Thread_queue_Extract( Thread_Control *the_thread ) { Thread_queue_Context queue_context; Thread_queue_Queue *queue; _Thread_queue_Context_initialize( &queue_context ); _Thread_queue_Context_clear_priority_updates( &queue_context ); _Thread_Wait_acquire( the_thread, &queue_context ); queue = the_thread->Wait.queue; if ( queue != NULL ) { bool unblock; _Thread_Wait_remove_request( the_thread, &queue_context.Lock_context ); _Thread_queue_Context_set_MP_callout( &queue_context, _Thread_queue_MP_callout_do_nothing ); unblock = _Thread_queue_Extract_locked( queue, the_thread->Wait.operations, the_thread, &queue_context ); _Thread_queue_Unblock_critical( unblock, queue, the_thread, &queue_context.Lock_context.Lock_context ); } else { _Thread_Wait_release( the_thread, &queue_context ); } } void _Thread_queue_Surrender( Thread_queue_Queue *queue, Thread_queue_Heads *heads, Thread_Control *previous_owner, Thread_queue_Context *queue_context, const Thread_queue_Operations *operations ) { Thread_Control *new_owner; bool unblock; Per_CPU_Control *cpu_self; _Assert( heads != NULL ); _Thread_queue_Context_clear_priority_updates( queue_context ); new_owner = ( *operations->surrender )( queue, heads, previous_owner, queue_context ); queue->owner = new_owner; #if defined(RTEMS_MULTIPROCESSING) if ( !_Thread_queue_MP_set_callout( new_owner, queue_context ) ) #endif { _Thread_Resource_count_increment( new_owner ); } unblock = _Thread_queue_Make_ready_again( new_owner ); cpu_self = _Thread_queue_Dispatch_disable( queue_context ); _Thread_queue_Queue_release( queue, &queue_context->Lock_context.Lock_context ); _Thread_Priority_update( queue_context ); if ( unblock ) { _Thread_Remove_timer_and_unblock( new_owner, queue ); } _Thread_Dispatch_enable( cpu_self ); } #if defined(RTEMS_SMP) void _Thread_queue_Surrender_sticky( Thread_queue_Queue *queue, Thread_queue_Heads *heads, Thread_Control *previous_owner, Thread_queue_Context *queue_context, const Thread_queue_Operations *operations ) { Thread_Control *new_owner; Per_CPU_Control *cpu_self; _Assert( heads != NULL ); _Thread_queue_Context_clear_priority_updates( queue_context ); new_owner = ( *operations->surrender )( queue, heads, previous_owner, queue_context ); queue->owner = new_owner; _Thread_queue_Make_ready_again( new_owner ); cpu_self = _Thread_queue_Dispatch_disable( queue_context ); _Thread_queue_Queue_release( queue, &queue_context->Lock_context.Lock_context ); _Thread_Priority_and_sticky_update( previous_owner, -1 ); _Thread_Priority_and_sticky_update( new_owner, 0 ); _Thread_Dispatch_enable( cpu_self ); } #endif Thread_Control *_Thread_queue_Do_dequeue( Thread_queue_Control *the_thread_queue, const Thread_queue_Operations *operations #if defined(RTEMS_MULTIPROCESSING) , Thread_queue_MP_callout mp_callout #endif ) { Thread_queue_Context queue_context; Thread_Control *the_thread; _Thread_queue_Context_initialize( &queue_context ); _Thread_queue_Context_set_MP_callout( &queue_context, mp_callout ); _Thread_queue_Acquire( the_thread_queue, &queue_context ); the_thread = _Thread_queue_First_locked( the_thread_queue, operations ); if ( the_thread != NULL ) { _Thread_queue_Extract_critical( &the_thread_queue->Queue, operations, the_thread, &queue_context ); } else { _Thread_queue_Release( the_thread_queue, &queue_context ); } return the_thread; } #if defined(RTEMS_MULTIPROCESSING) void _Thread_queue_Unblock_proxy( Thread_queue_Queue *queue, Thread_Control *the_thread ) { const Thread_queue_Object *the_queue_object; Thread_Proxy_control *the_proxy; Thread_queue_MP_callout mp_callout; the_queue_object = THREAD_QUEUE_QUEUE_TO_OBJECT( queue ); the_proxy = (Thread_Proxy_control *) the_thread; mp_callout = the_proxy->thread_queue_callout; ( *mp_callout )( the_thread, the_queue_object->Object.id ); _Thread_MP_Free_proxy( the_thread ); } #endif