diff options
author | Ralf Corsepius <ralf.corsepius@rtems.org> | 2009-11-29 13:51:53 +0000 |
---|---|---|
committer | Ralf Corsepius <ralf.corsepius@rtems.org> | 2009-11-29 13:51:53 +0000 |
commit | 28352faecf8b000b71f734cd728f38aa212b9669 (patch) | |
tree | c8e773b36bf32ba725cc1548e515f2fa9f8ebe96 /cpukit/score/src | |
parent | Whitespace removal. (diff) | |
download | rtems-28352faecf8b000b71f734cd728f38aa212b9669.tar.bz2 |
Whitespace removal.
Diffstat (limited to 'cpukit/score/src')
43 files changed, 98 insertions, 98 deletions
diff --git a/cpukit/score/src/apimutexlock.c b/cpukit/score/src/apimutexlock.c index ee74816100..6729478ae4 100644 --- a/cpukit/score/src/apimutexlock.c +++ b/cpukit/score/src/apimutexlock.c @@ -20,7 +20,7 @@ void _API_Mutex_Lock( API_Mutex_Control *the_mutex ) { - ISR_Level level; + ISR_Level level; _ISR_Disable( level ); diff --git a/cpukit/score/src/corebarrierwait.c b/cpukit/score/src/corebarrierwait.c index 9e71a8ac9a..4db276f1ce 100644 --- a/cpukit/score/src/corebarrierwait.c +++ b/cpukit/score/src/corebarrierwait.c @@ -60,14 +60,14 @@ void _CORE_barrier_Wait( _ISR_Disable( level ); the_barrier->number_of_waiting_threads++; if ( _CORE_barrier_Is_automatic( &the_barrier->Attributes ) ) { - if ( the_barrier->number_of_waiting_threads == + if ( the_barrier->number_of_waiting_threads == the_barrier->Attributes.maximum_count) { executing->Wait.return_code = CORE_BARRIER_STATUS_AUTOMATICALLY_RELEASED; _ISR_Enable( level ); _CORE_barrier_Release( the_barrier, id, api_barrier_mp_support ); return; } - } + } _Thread_queue_Enter_critical_section( &the_barrier->Wait_queue ); executing->Wait.queue = &the_barrier->Wait_queue; diff --git a/cpukit/score/src/coremsg.c b/cpukit/score/src/coremsg.c index ca5dfcfbeb..4e3b9545e1 100644 --- a/cpukit/score/src/coremsg.c +++ b/cpukit/score/src/coremsg.c @@ -62,9 +62,9 @@ bool _CORE_message_queue_Initialize( the_message_queue->number_of_pending_messages = 0; the_message_queue->maximum_message_size = maximum_message_size; _CORE_message_queue_Set_notify( the_message_queue, NULL, NULL ); - + /* - * Round size up to multiple of a pointer for chain init and + * Round size up to multiple of a pointer for chain init and * check for overflow on adding overhead to each message. */ allocated_message_size = maximum_message_size; @@ -72,29 +72,29 @@ bool _CORE_message_queue_Initialize( allocated_message_size += sizeof(uint32_t); allocated_message_size &= ~(sizeof(uint32_t) - 1); } - + if (allocated_message_size < maximum_message_size) return false; /* * Calculate how much total memory is required for message buffering and - * check for overflow on the multiplication. + * check for overflow on the multiplication. */ message_buffering_required = (size_t) maximum_pending_messages * (allocated_message_size + sizeof(CORE_message_queue_Buffer_control)); - + if (message_buffering_required < allocated_message_size) return false; /* * Attempt to allocate the message memory */ - the_message_queue->message_buffers = (CORE_message_queue_Buffer *) + the_message_queue->message_buffers = (CORE_message_queue_Buffer *) _Workspace_Allocate( message_buffering_required ); - + if (the_message_queue->message_buffers == 0) return false; - + /* * Initialize the pool of inactive messages, pending messages, * and set of waiting threads. @@ -105,9 +105,9 @@ bool _CORE_message_queue_Initialize( (size_t) maximum_pending_messages, allocated_message_size + sizeof( CORE_message_queue_Buffer_control ) ); - + _Chain_Initialize_empty( &the_message_queue->Pending_messages ); - + _Thread_queue_Initialize( &the_message_queue->Wait_queue, _CORE_message_queue_Is_priority( the_message_queue_attributes ) ? diff --git a/cpukit/score/src/coremsgseize.c b/cpukit/score/src/coremsgseize.c index 7b2294cb72..9a9fd75567 100644 --- a/cpukit/score/src/coremsgseize.c +++ b/cpukit/score/src/coremsgseize.c @@ -78,7 +78,7 @@ void _CORE_message_queue_Seize( _ISR_Enable( level ); *size_p = the_message->Contents.size; - _Thread_Executing->Wait.count = + _Thread_Executing->Wait.count = _CORE_message_queue_Get_message_priority( the_message ); _CORE_message_queue_Copy_buffer( the_message->Contents.buffer, @@ -88,7 +88,7 @@ void _CORE_message_queue_Seize( #if !defined(RTEMS_SCORE_COREMSG_ENABLE_BLOCKING_SEND) /* - * There is not an API with blocking sends enabled. + * There is not an API with blocking sends enabled. * So return immediately. */ _CORE_message_queue_Free_message_buffer(the_message_queue, the_message); @@ -119,7 +119,7 @@ void _CORE_message_queue_Seize( * waiting task. */ _CORE_message_queue_Set_message_priority( - the_message, + the_message, the_thread->Wait.count ); the_message->Contents.size = (size_t) the_thread->Wait.option; diff --git a/cpukit/score/src/coremsgsubmit.c b/cpukit/score/src/coremsgsubmit.c index a1cdae84ea..c67157dc44 100644 --- a/cpukit/score/src/coremsgsubmit.c +++ b/cpukit/score/src/coremsgsubmit.c @@ -108,7 +108,7 @@ CORE_message_queue_Status _CORE_message_queue_Submit( the_message = _CORE_message_queue_Allocate_message_buffer( the_message_queue ); - #if defined(RTEMS_DEBUG) + #if defined(RTEMS_DEBUG) /* * NOTE: If the system is consistent, this error should never occur. */ diff --git a/cpukit/score/src/coremutex.c b/cpukit/score/src/coremutex.c index 6d8a64e75d..dcd2e778fe 100644 --- a/cpukit/score/src/coremutex.c +++ b/cpukit/score/src/coremutex.c @@ -64,8 +64,8 @@ CORE_mutex_Status _CORE_mutex_Initialize( the_mutex->holder_id = _Thread_Executing->Object.id; if ( _CORE_mutex_Is_inherit_priority( &the_mutex->Attributes ) || _CORE_mutex_Is_priority_ceiling( &the_mutex->Attributes ) ) { - - if ( _Thread_Executing->current_priority < + + if ( _Thread_Executing->current_priority < the_mutex->Attributes.priority_ceiling ) return CORE_MUTEX_STATUS_CEILING_VIOLATED; #ifdef __RTEMS_STRICT_ORDER_MUTEX__ diff --git a/cpukit/score/src/coremutexsurrender.c b/cpukit/score/src/coremutexsurrender.c index 28d4b5d876..4d215bb2d1 100644 --- a/cpukit/score/src/coremutexsurrender.c +++ b/cpukit/score/src/coremutexsurrender.c @@ -84,7 +84,7 @@ CORE_mutex_Status _CORE_mutex_Surrender( the_mutex->nest_count--; - if ( the_mutex->nest_count != 0 ) { + if ( the_mutex->nest_count != 0 ) { /* * All error checking is on the locking side, so if the lock was * allowed to acquired multiple times, then we should just deal with diff --git a/cpukit/score/src/corerwlockobtainread.c b/cpukit/score/src/corerwlockobtainread.c index 59d984907c..78b6ebc187 100644 --- a/cpukit/score/src/corerwlockobtainread.c +++ b/cpukit/score/src/corerwlockobtainread.c @@ -26,7 +26,7 @@ * _CORE_rwlock_Obtain_for_reading * * This function waits for the rwlock to become available. Optionally, - * a limit may be placed on the duration of the spin. + * a limit may be placed on the duration of the spin. * * Input parameters: * the_rwlock - the rwlock control block to initialize diff --git a/cpukit/score/src/corerwlockobtainwrite.c b/cpukit/score/src/corerwlockobtainwrite.c index 339f257397..c3aae6e34c 100644 --- a/cpukit/score/src/corerwlockobtainwrite.c +++ b/cpukit/score/src/corerwlockobtainwrite.c @@ -26,7 +26,7 @@ * _CORE_rwlock_Obtain_for_writing * * This function waits for the rwlock to become available. Optionally, - * a limit may be placed on the duration of the spin. + * a limit may be placed on the duration of the spin. * * Input parameters: * the_rwlock - the rwlock control block to initialize diff --git a/cpukit/score/src/corerwlockrelease.c b/cpukit/score/src/corerwlockrelease.c index 98aa973580..379fdebdec 100644 --- a/cpukit/score/src/corerwlockrelease.c +++ b/cpukit/score/src/corerwlockrelease.c @@ -61,7 +61,7 @@ CORE_RWLock_Status _CORE_RWLock_Release( _ISR_Enable( level ); return CORE_RWLOCK_SUCCESSFUL; } - } + } /* CORE_RWLOCK_LOCKED_FOR_WRITING or READING with readers */ executing->Wait.return_code = CORE_RWLOCK_SUCCESSFUL; @@ -74,7 +74,7 @@ CORE_RWLock_Status _CORE_RWLock_Release( _ISR_Enable( level ); next = _Thread_queue_Dequeue( &the_rwlock->Wait_queue ); - + if ( next ) { if ( next->Wait.option == CORE_RWLOCK_THREAD_WAITING_FOR_WRITE ) { the_rwlock->current_state = CORE_RWLOCK_LOCKED_FOR_WRITING; @@ -92,7 +92,7 @@ CORE_RWLock_Status _CORE_RWLock_Release( */ while ( 1 ) { next = _Thread_queue_First( &the_rwlock->Wait_queue ); - if ( !next || + if ( !next || next->Wait.option == CORE_RWLOCK_THREAD_WAITING_FOR_WRITE ) return CORE_RWLOCK_SUCCESSFUL; the_rwlock->number_of_readers += 1; diff --git a/cpukit/score/src/coresemseize.c b/cpukit/score/src/coresemseize.c index 17f21a43c1..a66690e148 100644 --- a/cpukit/score/src/coresemseize.c +++ b/cpukit/score/src/coresemseize.c @@ -64,7 +64,7 @@ void _CORE_semaphore_Seize( return; } - /* + /* * If the semaphore was not available and the caller was not willing * to block, then return immediately with a status indicating that * the semaphore was not available and the caller never blocked. diff --git a/cpukit/score/src/corespinlockrelease.c b/cpukit/score/src/corespinlockrelease.c index 46aa7c1105..fe027448c1 100644 --- a/cpukit/score/src/corespinlockrelease.c +++ b/cpukit/score/src/corespinlockrelease.c @@ -50,7 +50,7 @@ CORE_spinlock_Status _CORE_spinlock_Release( if ( the_spinlock->lock == CORE_SPINLOCK_UNLOCKED ) { _ISR_Enable( level ); return CORE_SPINLOCK_NOT_LOCKED; - } + } /* * It must locked by the current thread before it can be unlocked. @@ -58,8 +58,8 @@ CORE_spinlock_Status _CORE_spinlock_Release( if ( the_spinlock->holder != _Thread_Executing->Object.id ) { _ISR_Enable( level ); return CORE_SPINLOCK_NOT_HOLDER; - } - + } + /* * Let it be unlocked. */ diff --git a/cpukit/score/src/corespinlockwait.c b/cpukit/score/src/corespinlockwait.c index 4fb49badd7..cf15163977 100644 --- a/cpukit/score/src/corespinlockwait.c +++ b/cpukit/score/src/corespinlockwait.c @@ -26,7 +26,7 @@ * _CORE_spinlock_Wait * * This function waits for the spinlock to become available. Optionally, - * a limit may be placed on the duration of the spin. + * a limit may be placed on the duration of the spin. * * Input parameters: * the_spinlock - the spinlock control block to initialize @@ -48,7 +48,7 @@ CORE_spinlock_Status _CORE_spinlock_Wait( #endif _ISR_Disable( level ); - if ( (the_spinlock->lock == CORE_SPINLOCK_LOCKED) && + if ( (the_spinlock->lock == CORE_SPINLOCK_LOCKED) && (the_spinlock->holder == _Thread_Executing->Object.id) ) { _ISR_Enable( level ); return CORE_SPINLOCK_HOLDER_RELOCKING; @@ -89,12 +89,12 @@ CORE_spinlock_Status _CORE_spinlock_Wait( * POSIX does not say anything about ISRs, that implies that * another thread must be able to run while spinning. We are * not blocking so that implies we are at least preemptible - * and possibly time-sliced. - * + * and possibly time-sliced. + * * So first, we will enable interrpts to allow for them to happen. * Then we will "flash" the thread dispatching critical section * so other threads have a chance to run. - * + * * A spinlock cannot be deleted while it is being used so we are * safe from deletion. */ @@ -104,7 +104,7 @@ CORE_spinlock_Status _CORE_spinlock_Wait( _Thread_Enable_dispatch(); /* Another thread could get dispatched here */ - + /* Reenter the critical sections so we can attempt the lock again. */ _Thread_Disable_dispatch(); diff --git a/cpukit/score/src/heapwalk.c b/cpukit/score/src/heapwalk.c index ee2ba0181f..f7043c0c41 100644 --- a/cpukit/score/src/heapwalk.c +++ b/cpukit/score/src/heapwalk.c @@ -366,7 +366,7 @@ bool _Heap_Walk( block, next_block ); - + return false; } @@ -378,7 +378,7 @@ bool _Heap_Walk( block, block_size ); - + return false; } @@ -391,7 +391,7 @@ bool _Heap_Walk( block_size, min_block_size ); - + return false; } @@ -403,10 +403,10 @@ bool _Heap_Walk( block, next_block ); - + return false; } - + if ( !_Heap_Is_prev_used( next_block ) ) { if ( !_Heap_Walk_check_free_block( source, printer, heap, block ) ) { return false; diff --git a/cpukit/score/src/mpci.c b/cpukit/score/src/mpci.c index 74ed371df5..d8f68ebc89 100644 --- a/cpukit/score/src/mpci.c +++ b/cpukit/score/src/mpci.c @@ -122,7 +122,7 @@ void _MPCI_Create_server( void ) &_Thread_Internal_information, _MPCI_Receive_server_tcb, NULL, /* allocate the stack */ - _Stack_Minimum() + + _Stack_Minimum() + CPU_MPCI_RECEIVE_SERVER_EXTRA_STACK + _Configuration_MP_table->extra_mpci_receive_server_stack, CPU_ALL_TASKS_ARE_FP, diff --git a/cpukit/score/src/objectextendinformation.c b/cpukit/score/src/objectextendinformation.c index bafdc58fda..f8e19cf130 100644 --- a/cpukit/score/src/objectextendinformation.c +++ b/cpukit/score/src/objectextendinformation.c @@ -94,7 +94,7 @@ void _Objects_Extend_information( * Allocate the name table, and the objects and if it fails either return or * generate a fatal error depending on auto-extending being active. */ - + block_size = information->allocation_size * information->size; if ( information->auto_extend ) { new_object_block = _Workspace_Allocate( block_size ); @@ -103,7 +103,7 @@ void _Objects_Extend_information( } else { new_object_block = _Workspace_Allocate_or_fatal_error( block_size ); } - + /* * If the index_base is the maximum we need to grow the tables. */ @@ -152,7 +152,7 @@ void _Objects_Extend_information( _Workspace_Free( new_object_block ); return; } - + /* * Break the block into the various sections. */ diff --git a/cpukit/score/src/objectget.c b/cpukit/score/src/objectget.c index a0aa3465b2..2abcc389fb 100644 --- a/cpukit/score/src/objectget.c +++ b/cpukit/score/src/objectget.c @@ -63,7 +63,7 @@ Objects_Control *_Objects_Get( * outside this class. * * If the Id matches the api, class, and node but index portion is 0, - * then the subtraction will underflow and the addition of 1 will + * then the subtraction will underflow and the addition of 1 will * result in a 0 index. The zeroth element in the local_table is * always NULL. * diff --git a/cpukit/score/src/pheapresizeblock.c b/cpukit/score/src/pheapresizeblock.c index c229f31731..14dc522cff 100644 --- a/cpukit/score/src/pheapresizeblock.c +++ b/cpukit/score/src/pheapresizeblock.c @@ -35,7 +35,7 @@ bool _Protected_heap_Resize_block( uintptr_t avail_mem_size; _RTEMS_Lock_allocator(); - status = _Heap_Resize_block( + status = _Heap_Resize_block( the_heap, starting_address, size, &old_mem_size, &avail_mem_size ); _RTEMS_Unlock_allocator(); return (status == HEAP_RESIZE_SUCCESSFUL); diff --git a/cpukit/score/src/pheapwalk.c b/cpukit/score/src/pheapwalk.c index e86874192f..4aa2279902 100644 --- a/cpukit/score/src/pheapwalk.c +++ b/cpukit/score/src/pheapwalk.c @@ -36,7 +36,7 @@ bool _Protected_heap_Walk( * If we are called from within a dispatching critical section, * then it is forbidden to lock a mutex. But since we are inside * a critical section, it should be safe to walk it unlocked. - * + * * NOTE: Dispatching is also disabled during initialization. */ if ( !_Thread_Dispatch_disable_level ) { diff --git a/cpukit/score/src/threadclearstate.c b/cpukit/score/src/threadclearstate.c index 9049c8bbee..872d57cc66 100644 --- a/cpukit/score/src/threadclearstate.c +++ b/cpukit/score/src/threadclearstate.c @@ -78,7 +78,7 @@ void _Thread_Clear_state( * then we have a new heir. This may or may not result in a * context switch. * - * Normal case: + * Normal case: * If the current thread is preemptible, then we need to do * a context switch. * Pseudo-ISR case: diff --git a/cpukit/score/src/threadget.c b/cpukit/score/src/threadget.c index d7d9106998..c531fd9bf0 100644 --- a/cpukit/score/src/threadget.c +++ b/cpukit/score/src/threadget.c @@ -56,26 +56,26 @@ Thread_Control *_Thread_Get ( Objects_Information **api_information; Objects_Information *information; Thread_Control *tp = (Thread_Control *) 0; - + if ( _Objects_Are_ids_equal( id, OBJECTS_ID_OF_SELF ) ) { _Thread_Disable_dispatch(); *location = OBJECTS_LOCAL; tp = _Thread_Executing; goto done; } - + the_api = _Objects_Get_API( id ); if ( !_Objects_Is_api_valid( the_api ) ) { *location = OBJECTS_ERROR; goto done; } - + the_class = _Objects_Get_class( id ); if ( the_class != 1 ) { /* threads are always first class :) */ *location = OBJECTS_ERROR; goto done; } - + api_information = _Objects_Information_table[ the_api ]; if ( !api_information ) { *location = OBJECTS_ERROR; @@ -87,9 +87,9 @@ Thread_Control *_Thread_Get ( *location = OBJECTS_ERROR; goto done; } - + tp = (Thread_Control *) _Objects_Get( information, id, location ); - + done: return tp; } diff --git a/cpukit/score/src/threadhandler.c b/cpukit/score/src/threadhandler.c index be073700a0..3fffd8fb83 100644 --- a/cpukit/score/src/threadhandler.c +++ b/cpukit/score/src/threadhandler.c @@ -88,7 +88,7 @@ void _Thread_Handler( void ) executing = _Thread_Executing; /* - * Some CPUs need to tinker with the call frame or registers when the + * Some CPUs need to tinker with the call frame or registers when the * thread actually begins to execute for the first time. This is a * hook point where the port gets a shot at doing whatever it requires. */ diff --git a/cpukit/score/src/threadinitialize.c b/cpukit/score/src/threadinitialize.c index 2f7cf370be..fe9a5c3dea 100644 --- a/cpukit/score/src/threadinitialize.c +++ b/cpukit/score/src/threadinitialize.c @@ -214,7 +214,7 @@ bool _Thread_Initialize( /* * We assume the Allocator Mutex is locked and dispatching is * enabled when we get here. We want to be able to run the - * user extensions with dispatching enabled. The Allocator + * user extensions with dispatching enabled. The Allocator * Mutex provides sufficient protection to let the user extensions * run safely. */ @@ -229,7 +229,7 @@ failed: for ( i=0 ; i <= THREAD_API_LAST ; i++ ) if ( the_thread->API_Extensions[i] ) _Workspace_Free( the_thread->API_Extensions[i] ); - + if ( extensions_area ) (void) _Workspace_Free( extensions_area ); diff --git a/cpukit/score/src/threadqenqueue.c b/cpukit/score/src/threadqenqueue.c index b517cc1090..dbb1661feb 100644 --- a/cpukit/score/src/threadqenqueue.c +++ b/cpukit/score/src/threadqenqueue.c @@ -71,7 +71,7 @@ void _Thread_queue_Enqueue_with_handler( /* * If the thread wants to timeout, then schedule its timer. - */ + */ if ( timeout ) { _Watchdog_Initialize( &the_thread->Timer, @@ -85,7 +85,7 @@ void _Thread_queue_Enqueue_with_handler( /* * Now enqueue the thread per the discipline for this thread queue. - */ + */ if ( the_thread_queue->discipline == THREAD_QUEUE_DISCIPLINE_PRIORITY ) enqueue_p = _Thread_queue_Enqueue_priority; else /* must be THREAD_QUEUE_DISCIPLINE_FIFO */ diff --git a/cpukit/score/src/threadqenqueuefifo.c b/cpukit/score/src/threadqenqueuefifo.c index 229261efad..da3c809212 100644 --- a/cpukit/score/src/threadqenqueuefifo.c +++ b/cpukit/score/src/threadqenqueuefifo.c @@ -65,7 +65,7 @@ Thread_blocking_operation_States _Thread_queue_Enqueue_fifo ( _ISR_Enable( level ); return THREAD_BLOCKING_OPERATION_NOTHING_HAPPENED; } - + /* * An interrupt completed the thread's blocking request. * For example, the blocking thread could have been given diff --git a/cpukit/score/src/timespecaddto.c b/cpukit/score/src/timespecaddto.c index e45e1e1d7b..ac764b83c1 100644 --- a/cpukit/score/src/timespecaddto.c +++ b/cpukit/score/src/timespecaddto.c @@ -1,4 +1,4 @@ -/** +/** * @file score/src/timespecaddto.c */ diff --git a/cpukit/score/src/timespecdivide.c b/cpukit/score/src/timespecdivide.c index 68b284909d..5e2cd2a65a 100644 --- a/cpukit/score/src/timespecdivide.c +++ b/cpukit/score/src/timespecdivide.c @@ -1,4 +1,4 @@ -/** +/** * @file score/src/timespecdivide.c */ diff --git a/cpukit/score/src/timespecdividebyinteger.c b/cpukit/score/src/timespecdividebyinteger.c index 34a719a577..118fd24ee3 100644 --- a/cpukit/score/src/timespecdividebyinteger.c +++ b/cpukit/score/src/timespecdividebyinteger.c @@ -1,4 +1,4 @@ -/** +/** * @file score/src/timespecdividebyinteger.c */ diff --git a/cpukit/score/src/timespecgreaterthan.c b/cpukit/score/src/timespecgreaterthan.c index 3db4c7c877..b0e7870ac6 100644 --- a/cpukit/score/src/timespecgreaterthan.c +++ b/cpukit/score/src/timespecgreaterthan.c @@ -1,4 +1,4 @@ -/** +/** * @file score/src/timespecgreaterthan.c */ @@ -36,7 +36,7 @@ bool _Timespec_Greater_than( /* ASSERT: lhs->tv_sec == rhs->tv_sec */ if ( lhs->tv_nsec > rhs->tv_nsec ) - return true; + return true; return false; } diff --git a/cpukit/score/src/timespecisvalid.c b/cpukit/score/src/timespecisvalid.c index c02e8917ca..3347052ba9 100644 --- a/cpukit/score/src/timespecisvalid.c +++ b/cpukit/score/src/timespecisvalid.c @@ -1,4 +1,4 @@ -/** +/** * @file score/src/timespecisvalid.c */ diff --git a/cpukit/score/src/timespeclessthan.c b/cpukit/score/src/timespeclessthan.c index 9b9000bdbb..fb54e69dc9 100644 --- a/cpukit/score/src/timespeclessthan.c +++ b/cpukit/score/src/timespeclessthan.c @@ -1,4 +1,4 @@ -/** +/** * @file score/src/timespeclessthan.c */ @@ -36,7 +36,7 @@ bool _Timespec_Less_than( /* ASSERT: lhs->tv_sec == rhs->tv_sec */ if ( lhs->tv_nsec < rhs->tv_nsec ) - return true; + return true; return false; } diff --git a/cpukit/score/src/timespecsubtract.c b/cpukit/score/src/timespecsubtract.c index d17107fcc0..44e9f21c25 100644 --- a/cpukit/score/src/timespecsubtract.c +++ b/cpukit/score/src/timespecsubtract.c @@ -1,4 +1,4 @@ -/** +/** * @file score/src/timespecsubtract.c */ @@ -32,7 +32,7 @@ void _Timespec_Subtract( if (end->tv_nsec < start->tv_nsec) { result->tv_sec = end->tv_sec - start->tv_sec - 1; - result->tv_nsec = + result->tv_nsec = (TOD_NANOSECONDS_PER_SECOND - start->tv_nsec) + end->tv_nsec; } else { result->tv_sec = end->tv_sec - start->tv_sec; diff --git a/cpukit/score/src/timespectoticks.c b/cpukit/score/src/timespectoticks.c index 51ad0e07f3..ddd82e67a7 100644 --- a/cpukit/score/src/timespectoticks.c +++ b/cpukit/score/src/timespectoticks.c @@ -1,4 +1,4 @@ -/** +/** * @file score/src/timespectoticks.c */ @@ -36,7 +36,7 @@ uint32_t _Timespec_To_ticks( { uint32_t ticks; - if ( (time->tv_sec == 0) && (time->tv_nsec == 0) ) + if ( (time->tv_sec == 0) && (time->tv_nsec == 0) ) return 0; ticks = time->tv_sec * TOD_TICKS_PER_SECOND; diff --git a/cpukit/score/src/ts64divide.c b/cpukit/score/src/ts64divide.c index 0bf410664b..01e2661ac4 100644 --- a/cpukit/score/src/ts64divide.c +++ b/cpukit/score/src/ts64divide.c @@ -1,4 +1,4 @@ -/** +/** * @file score/src/ts64divide.c */ diff --git a/cpukit/score/src/ts64getnanoseconds.c b/cpukit/score/src/ts64getnanoseconds.c index 86ba1c1b75..f9d57a9ab1 100644 --- a/cpukit/score/src/ts64getnanoseconds.c +++ b/cpukit/score/src/ts64getnanoseconds.c @@ -1,4 +1,4 @@ -/** +/** * @file score/src/ts64toticks.c */ diff --git a/cpukit/score/src/ts64toticks.c b/cpukit/score/src/ts64toticks.c index d751a109a1..273bcc15b4 100644 --- a/cpukit/score/src/ts64toticks.c +++ b/cpukit/score/src/ts64toticks.c @@ -1,4 +1,4 @@ -/** +/** * @file score/src/ts64toticks.c */ diff --git a/cpukit/score/src/userext.c b/cpukit/score/src/userext.c index 9b778b6300..d0e8efc6c9 100644 --- a/cpukit/score/src/userext.c +++ b/cpukit/score/src/userext.c @@ -45,13 +45,13 @@ void _User_extensions_Handler_initialization(void) _Workspace_Allocate_or_fatal_error( number_of_extensions * sizeof( User_extensions_Control ) ); - + memset ( extension, 0, number_of_extensions * sizeof( User_extensions_Control ) ); - + for ( i = 0 ; i < number_of_extensions ; i++ ) { _User_extensions_Add_set_with_table (extension, &initial_extensions[i]); extension++; diff --git a/cpukit/score/src/userextaddapiset.c b/cpukit/score/src/userextaddapiset.c index 5b54e19069..954571f8af 100644 --- a/cpukit/score/src/userextaddapiset.c +++ b/cpukit/score/src/userextaddapiset.c @@ -36,7 +36,7 @@ void _User_extensions_Add_API_set ( #endif { the_extension->Switch.thread_switch = the_extension->Callouts.thread_switch; - _Chain_Append( + _Chain_Append( &_User_extensions_Switches_list, &the_extension->Switch.Node ); } } diff --git a/cpukit/score/src/userextremoveset.c b/cpukit/score/src/userextremoveset.c index 4466007c25..a7f23da536 100644 --- a/cpukit/score/src/userextremoveset.c +++ b/cpukit/score/src/userextremoveset.c @@ -29,7 +29,7 @@ void _User_extensions_Remove_set ( ) { _Chain_Extract( &the_extension->Node ); - + /* * If a switch handler is present, remove it. */ diff --git a/cpukit/score/src/userextthreadswitch.c b/cpukit/score/src/userextthreadswitch.c index e7896e79a0..10ab6b6583 100644 --- a/cpukit/score/src/userextthreadswitch.c +++ b/cpukit/score/src/userextthreadswitch.c @@ -31,7 +31,7 @@ void _User_extensions_Thread_switch ( { Chain_Node *the_node; User_extensions_Switch_control *the_extension_switch; - + for ( the_node = _User_extensions_Switches_list.first ; !_Chain_Is_tail( &_User_extensions_Switches_list, the_node ) ; the_node = the_node->next ) { diff --git a/cpukit/score/src/watchdogadjusttochain.c b/cpukit/score/src/watchdogadjusttochain.c index 0578276429..569a9364f7 100644 --- a/cpukit/score/src/watchdogadjusttochain.c +++ b/cpukit/score/src/watchdogadjusttochain.c @@ -31,12 +31,12 @@ void _Watchdog_Adjust_to_chain( { Watchdog_Interval units = units_arg; ISR_Level level; - Watchdog_Control *first; - + Watchdog_Control *first; + if ( units <= 0 ) { return; } - + _ISR_Disable( level ); while ( 1 ) { @@ -69,7 +69,7 @@ void _Watchdog_Adjust_to_chain( _Chain_Append_unprotected( to_fire, &first->Node ); _ISR_Flash( level ); - + if ( _Chain_Is_empty( header ) ) break; first = _Watchdog_First( header ); diff --git a/cpukit/score/src/watchdogreport.c b/cpukit/score/src/watchdogreport.c index fb1885aa4e..58bce56959 100644 --- a/cpukit/score/src/watchdogreport.c +++ b/cpukit/score/src/watchdogreport.c @@ -31,8 +31,8 @@ void _Watchdog_Report( "%s%s%4d %5d %p %p 0x%08x %p\n", ((name) ? name : ""), ((name) ? " " : ""), - watch->delta_interval, - watch->initial, + watch->delta_interval, + watch->initial, watch, watch->routine, watch->id, diff --git a/cpukit/score/src/watchdogtickle.c b/cpukit/score/src/watchdogtickle.c index e5482e7ca2..09b3c7bf09 100644 --- a/cpukit/score/src/watchdogtickle.c +++ b/cpukit/score/src/watchdogtickle.c @@ -53,37 +53,37 @@ void _Watchdog_Tickle( goto leave; the_watchdog = _Watchdog_First( header ); - + /* * For some reason, on rare occasions the_watchdog->delta_interval * of the head of the watchdog chain is 0. Before this test was * added, on these occasions an event (which usually was supposed * to have a timeout of 1 tick would have a delta_interval of 0, which - * would be decremented to 0xFFFFFFFF by the unprotected + * would be decremented to 0xFFFFFFFF by the unprotected * "the_watchdog->delta_interval--;" operation. * This would mean the event would not timeout, and also the chain would * be blocked, because a timeout with a very high number would be at the * head, rather than at the end. * The test "if (the_watchdog->delta_interval != 0)" - * here prevents this from occuring. - * - * We were not able to categorically identify the situation that causes - * this, but proved it to be true empirically. So this check causes + * here prevents this from occuring. + * + * We were not able to categorically identify the situation that causes + * this, but proved it to be true empirically. So this check causes * correct behaviour in this circumstance. - * + * * The belief is that a race condition exists whereby an event at the head * of the chain is removed (by a pending ISR or higher priority task) - * during the _ISR_Flash( level ); in _Watchdog_Insert, but the watchdog - * to be inserted has already had its delta_interval adjusted to 0, and - * so is added to the head of the chain with a delta_interval of 0. - * + * during the _ISR_Flash( level ); in _Watchdog_Insert, but the watchdog + * to be inserted has already had its delta_interval adjusted to 0, and + * so is added to the head of the chain with a delta_interval of 0. + * * Steven Johnson - 12/2005 (gcc-3.2.3 -O3 on powerpc) */ if (the_watchdog->delta_interval != 0) { the_watchdog->delta_interval--; if ( the_watchdog->delta_interval != 0 ) goto leave; - } + } do { watchdog_state = _Watchdog_Remove( the_watchdog ); |