(file) Return to Thread.cpp CVS log (file) (dir) Up to [Pegasus] / pegasus / src / Pegasus / Common

Diff for /pegasus/src/Pegasus/Common/Thread.cpp between version 1.36.4.6 and 1.56

version 1.36.4.6, 2003/08/14 14:26:20 version 1.56, 2003/10/22 14:26:04
Line 1 
Line 1 
 //%/////////////////////////////////////////////////////////////////////////////  //%2003////////////////////////////////////////////////////////////////////////
 // //
 // Copyright (c) 2000, 2001, 2002 BMC Software, Hewlett-Packard Company, IBM,  // Copyright (c) 2000, 2001, 2002  BMC Software, Hewlett-Packard Development
 // The Open Group, Tivoli Systems  // Company, L. P., IBM Corp., The Open Group, Tivoli Systems.
   // Copyright (c) 2003 BMC Software; Hewlett-Packard Development Company, L. P.;
   // IBM Corp.; EMC Corporation, The Open Group.
 // //
 // Permission is hereby granted, free of charge, to any person obtaining a copy // Permission is hereby granted, free of charge, to any person obtaining a copy
 // of this software and associated documentation files (the "Software"), to // of this software and associated documentation files (the "Software"), to
Line 51 
Line 53 
       ::operator delete(data);       ::operator delete(data);
 } }
  
 Boolean Thread::_signals_blocked = false;  // l10n start
 // l10n  void language_delete(void * data)
   {
      if( data != NULL)
      {
         AcceptLanguages * al = static_cast<AcceptLanguages *>(data);
         delete al;
      }
   }
   // l10n end
  
   Boolean Thread::_signals_blocked = false;
 // l10n // l10n
 PEGASUS_THREAD_KEY_TYPE Thread::_platform_thread_key;  PEGASUS_THREAD_KEY_TYPE Thread::_platform_thread_key = -1;
 Boolean Thread::_key_initialized = false; Boolean Thread::_key_initialized = false;
 Boolean Thread::_key_error = false; Boolean Thread::_key_error = false;
  
   
 // for non-native implementations // for non-native implementations
 #ifndef PEGASUS_THREAD_CLEANUP_NATIVE #ifndef PEGASUS_THREAD_CLEANUP_NATIVE
 void Thread::cleanup_push( void (*routine)(void *), void *parm) throw(IPCException) void Thread::cleanup_push( void (*routine)(void *), void *parm) throw(IPCException)
Line 209 
Line 220 
    {    {
                 // deletes the old tsd and creates a new one                 // deletes the old tsd and creates a new one
                 currentThrd->put_tsd("acceptLanguages",                 currentThrd->put_tsd("acceptLanguages",
                         thread_data::default_delete,                          language_delete,
                         sizeof(AcceptLanguages *),                         sizeof(AcceptLanguages *),
                         langs);                         langs);
    }    }
Line 232 
Line 243 
 } }
 // l10n end // l10n end
  
   
   
   // two special synchronization classes for ThreadPool
   //
   
   class timed_mutex
   {
      public:
         timed_mutex(Mutex* mut, int msec)
            :_mut(mut)
         {
            _mut->timed_lock(msec, pegasus_thread_self());
         }
         ~timed_mutex(void)
         {
            _mut->unlock();
         }
         Mutex* _mut;
   };
   
   
   class try_mutex
   {
      public:
         try_mutex(Mutex* mut)
            :_mut(mut)
         {
            _mut->try_lock(pegasus_thread_self());
         }
         ~try_mutex(void)
         {
            _mut->unlock();
         }
   
         Mutex* _mut;
   };
   
   
 DQueue<ThreadPool> ThreadPool::_pools(true); DQueue<ThreadPool> ThreadPool::_pools(true);
  
  
Line 296 
Line 345 
 } }
  
  
   // Note:   <<< Fri Oct 17 09:19:03 2003 mdd >>>
   // the pegasus_yield() calls that preceed each th->join() are to
   // give a thread on the running list a chance to reach a cancellation
   // point before the join
  
 ThreadPool::~ThreadPool(void) ThreadPool::~ThreadPool(void)
 { {
    try    try
    {    {
       _pools.remove(this);        // set the dying flag so all thread know the destructor has been entered
         {
            auto_mutex(&(this->_monitor));
       _dying++;       _dying++;
         }
         // remove from the global pools list
         _pools.remove(this);
   
         // start with idle threads.
       Thread *th = 0;       Thread *th = 0;
       th = _pool.remove_first();       th = _pool.remove_first();
         Semaphore* sleep_sem;
   
       while(th != 0)       while(th != 0)
       {       {
          Semaphore *sleep_sem = (Semaphore *)th->reference_tsd("sleep sem");           sleep_sem = (Semaphore *)th->reference_tsd("sleep sem");
   
          if(sleep_sem == 0)          if(sleep_sem == 0)
          {          {
             th->dereference_tsd();             th->dereference_tsd();
             throw NullPointer();             throw NullPointer();
          }          }
  
            // Signal to get the thread out of the work loop.
          sleep_sem->signal();          sleep_sem->signal();
   
            // Signal to get the thread past the end. See the comment
            // "wait to be awakend by the thread pool destructor"
            // Note: the current implementation of Thread for Windows
            // does not implement "pthread" cancelation points so this
            // is needed.
          sleep_sem->signal();          sleep_sem->signal();
          th->dereference_tsd();          th->dereference_tsd();
          // signal the thread's sleep semaphore  
          th->cancel();          th->cancel();
          th->join();          th->join();
          th->empty_tsd();  
          delete th;          delete th;
          th = _pool.remove_first();          th = _pool.remove_first();
       }       }
         th = _dead.remove_first();
       th = _running.remove_first();  
       while(th != 0)       while(th != 0)
       {       {
            sleep_sem = (Semaphore *)th->reference_tsd("sleep sem");
   
            if(sleep_sem == 0)
            {
               th->dereference_tsd();
               throw NullPointer();
            }
   
          // signal the thread's sleep semaphore          // signal the thread's sleep semaphore
            sleep_sem->signal();
            sleep_sem->signal();
            th->dereference_tsd();
          th->cancel();          th->cancel();
          th->join();          th->join();
          th->empty_tsd();  
          delete th;          delete th;
          th = _running.remove_first();           th = _dead.remove_first();
       }       }
  
       th = _dead.remove_first();        {
            th = _running.remove_first();
       while(th != 0)       while(th != 0)
       {       {
          // signal the thread's sleep semaphore          // signal the thread's sleep semaphore
   
               sleep_sem = (Semaphore *)th->reference_tsd("sleep sem");
               if(sleep_sem == 0 )
               {
                  th->dereference_tsd();
                  throw NullPointer();
               }
   
               sleep_sem->signal();
               sleep_sem->signal();
               th->dereference_tsd();
          th->cancel();          th->cancel();
               pegasus_yield();
   
          th->join();          th->join();
          th->empty_tsd();  
          delete th;          delete th;
          th = _dead.remove_first();              th = _running.remove_first();
            }
       }       }
   
    }    }
   
    catch(...)    catch(...)
    {    {
    }    }
Line 376 
Line 467 
       PEG_METHOD_EXIT();       PEG_METHOD_EXIT();
       throw NullPointer();       throw NullPointer();
    }    }
      if(pool->_dying.value())
      {
         PEG_METHOD_EXIT();
         return((PEGASUS_THREAD_RETURN)0);
      }
   
    Semaphore *sleep_sem = 0;    Semaphore *sleep_sem = 0;
    Semaphore *blocking_sem = 0;    Semaphore *blocking_sem = 0;
  
Line 388 
Line 485 
       deadlock_timer = (struct timeval *)myself->reference_tsd("deadlock timer");       deadlock_timer = (struct timeval *)myself->reference_tsd("deadlock timer");
       myself->dereference_tsd();       myself->dereference_tsd();
    }    }
    catch(IPCException &)  
    {  
       PEG_METHOD_EXIT();  
       myself->exit_self(0);  
    }  
    catch(...)    catch(...)
    {    {
       PEG_METHOD_EXIT();       PEG_METHOD_EXIT();
       myself->exit_self(0);        return((PEGASUS_THREAD_RETURN)0);
    }    }
  
    if(sleep_sem == 0 || deadlock_timer == 0)    if(sleep_sem == 0 || deadlock_timer == 0)
    {    {
       PEG_METHOD_EXIT();       PEG_METHOD_EXIT();
       throw NullPointer();        return((PEGASUS_THREAD_RETURN)0);
    }    }
  
    while(pool->_dying < 1)     while(1)
      {
         try
         {
            if(pool->_dying.value())
               break;
         }
         catch(...)
         {
            return((PEGASUS_THREAD_RETURN)0);
         }
   
         try
    {    {
       sleep_sem->wait();       sleep_sem->wait();
         }
         catch(IPCException& )
         {
            PEG_METHOD_EXIT();
            return((PEGASUS_THREAD_RETURN)0);
         }
  
       // when we awaken we reside on the running queue, not the pool queue       // when we awaken we reside on the running queue, not the pool queue
       if(pool->_dying > 0)        if(pool->_dying.value())
          break;        {
            PEG_METHOD_EXIT();
            return((PEGASUS_THREAD_RETURN)0);
         }
   
  
       PEGASUS_THREAD_RETURN (PEGASUS_THREAD_CDECL *_work)(void *) = 0;       PEGASUS_THREAD_RETURN (PEGASUS_THREAD_CDECL *_work)(void *) = 0;
       void *parm = 0;       void *parm = 0;
Line 430 
Line 545 
       catch(IPCException &)       catch(IPCException &)
       {       {
          PEG_METHOD_EXIT();          PEG_METHOD_EXIT();
          myself->exit_self(0);           return((PEGASUS_THREAD_RETURN)0);
       }       }
  
       if(_work == 0)       if(_work == 0)
Line 448 
Line 563 
       gettimeofday(deadlock_timer, NULL);       gettimeofday(deadlock_timer, NULL);
       try       try
       {       {
            {
               timed_mutex(&(pool->_monitor), 1000);
               if(pool->_dying.value())
               {
                  _undertaker(parm);
               }
            }
          _work(parm);          _work(parm);
       }       }
       catch(...)       catch(...)
       {       {
          gettimeofday(deadlock_timer, NULL);           return((PEGASUS_THREAD_RETURN)0);
       }       }
  
         // put myself back onto the available list
         try
         {
            timed_mutex(&(pool->_monitor), 1000);
            if(pool->_dying.value() == 0)
            {
       gettimeofday(deadlock_timer, NULL);       gettimeofday(deadlock_timer, NULL);
       if( blocking_sem != 0 )       if( blocking_sem != 0 )
          blocking_sem->signal();          blocking_sem->signal();
  
       // put myself back onto the available list              // If we are not on _running then ~ThreadPool has removed
       try              // us and now "owns" our pointer.
               if( pool->_running.remove((void *)myself) != 0 )
                   pool->_pool.insert_first(myself);
               else
       {       {
          pool->_running.remove((void *)myself);                 return((PEGASUS_THREAD_RETURN)0);
          pool->_link_pool(myself);  
       }       }
       catch(IPCException &)  
            }
            else
            {
               PEG_METHOD_EXIT();
               return((PEGASUS_THREAD_RETURN)0);
            }
         }
         catch(...)
       {       {
          PEG_METHOD_EXIT();          PEG_METHOD_EXIT();
          myself->exit_self(0);           return((PEGASUS_THREAD_RETURN)0);
       }       }
   
    }    }
   
      // TODO: Why is this needed? Why not just continue?
    // wait to be awakend by the thread pool destructor    // wait to be awakend by the thread pool destructor
    sleep_sem->wait();     //sleep_sem->wait();
   
    myself->test_cancel();    myself->test_cancel();
  
    PEG_METHOD_EXIT();    PEG_METHOD_EXIT();
    myself->exit_self(0);  
    return((PEGASUS_THREAD_RETURN)0);    return((PEGASUS_THREAD_RETURN)0);
 } }
  
Line 490 
Line 631 
    PEG_METHOD_ENTER(TRC_THREAD, "ThreadPool::allocate_and_awaken");    PEG_METHOD_ENTER(TRC_THREAD, "ThreadPool::allocate_and_awaken");
    struct timeval start;    struct timeval start;
    gettimeofday(&start, NULL);    gettimeofday(&start, NULL);
      Thread *th = 0;
   
      try
      {
         timed_mutex(&(this->_monitor), 1000);
         if(_dying.value())
         {
            return;
         }
         th = _pool.remove_first();
      }
      catch(...)
      {
         return;
   
      }
  
    Thread *th = _pool.remove_first();  
  
    // wait for the right interval and try again    // wait for the right interval and try again
    while (th == 0 && _dying < 1)     while (th == 0 && _dying.value() < 1)
    {    {
         // will throw an IPCException&
       _check_deadlock(&start) ;       _check_deadlock(&start) ;
  
       if(_max_threads == 0 || _current_threads < _max_threads)       if(_max_threads == 0 || _current_threads < _max_threads)
Line 504 
Line 661 
          continue;          continue;
       }       }
       pegasus_yield();       pegasus_yield();
         try
         {
            timed_mutex(&(this->_monitor), 1000);
            if(_dying.value())
            {
               return;
            }
       th = _pool.remove_first();       th = _pool.remove_first();
    }    }
         catch(...)
         {
            return ;
         }
      }
  
      if(_dying.value() < 1)
    if(_dying < 1)  
    {    {
       // initialize the thread data with the work function and parameters       // initialize the thread data with the work function and parameters
       Tracer::trace(TRC_THREAD, Tracer::LEVEL4,       Tracer::trace(TRC_THREAD, Tracer::LEVEL4,
Line 524 
Line 692 
       th->delete_tsd("blocking sem");       th->delete_tsd("blocking sem");
       if(blocking != 0 )       if(blocking != 0 )
          th->put_tsd("blocking sem", NULL, sizeof(Semaphore *), blocking);          th->put_tsd("blocking sem", NULL, sizeof(Semaphore *), blocking);
         try
         {
            timed_mutex(&(this->_monitor), 1000);
            if(_dying.value())
            {
               th->cancel();
               th->join();
               delete th;
               return;
            }
  
       // put the thread on the running list       // put the thread on the running list
       _running.insert_first(th);  
  
   
            _running.insert_first(th);
       // signal the thread's sleep semaphore to awaken it       // signal the thread's sleep semaphore to awaken it
       Semaphore *sleep_sem = (Semaphore *)th->reference_tsd("sleep sem");       Semaphore *sleep_sem = (Semaphore *)th->reference_tsd("sleep sem");
  
Line 541 
Line 720 
       sleep_sem->signal();       sleep_sem->signal();
       th->dereference_tsd();       th->dereference_tsd();
    }    }
         catch(...)
         {
            PEG_METHOD_EXIT();
            return;
         }
   
      }
    else    else
       _pool.insert_first(th);     {
         th->cancel();
         th->join();
         delete th;
      }
  
    PEG_METHOD_EXIT();    PEG_METHOD_EXIT();
 } }
Line 559 
Line 749 
    Uint32 bodies = 0;    Uint32 bodies = 0;
  
    // first go thread the dead q and clean it up as much as possible    // first go thread the dead q and clean it up as much as possible
    while(_dead.count() > 0)     try
      {
         timed_mutex(&(this->_monitor), 1000);
         if(_dying.value() )
         {
            return 0;
         }
   
         while(_dead.count() > 0 && _dying.value() == 0 )
    {    {
       Tracer::trace(TRC_THREAD, Tracer::LEVEL4, "ThreadPool:: removing and joining dead thread");       Tracer::trace(TRC_THREAD, Tracer::LEVEL4, "ThreadPool:: removing and joining dead thread");
       Thread *dead = _dead.remove_first();       Thread *dead = _dead.remove_first();
   
       if(dead == 0)       if(dead == 0)
          throw NullPointer();          throw NullPointer();
       dead->join();       dead->join();
       delete dead;       delete dead;
    }    }
      }
      catch(...)
      {
      }
   
  
    DQueue<Thread> * map[2] =    DQueue<Thread> * map[2] =
       {       {
Line 598 
Line 802 
    for( ; i < 2; i++)    for( ; i < 2; i++)
 #endif #endif
    {    {
         try
         {
            try_mutex(&(this->_monitor));
         }
         catch(IPCException&)
         {
            return bodies;
         }
   
       q = map[i];       q = map[i];
       if(q->count() > 0 )       if(q->count() > 0 )
       {       {
          try          try
          {          {
               if(_dying.value())
               {
                  return bodies;
               }
   
             q->try_lock();             q->try_lock();
          }          }
          catch(...)          catch(...)
Line 703 
Line 921 
                }                }
             }             }
             th = q->next(th);             th = q->next(th);
             pegasus_sleep(1);              pegasus_yield();
          }          }
          q->unlock();          q->unlock();
          while (needed.value() > 0)        }
          {     }
      if(_dying.value() )
         return bodies;
   
      while (needed.value() > 0)   {
             _link_pool(_init_thread());             _link_pool(_init_thread());
             needed--;             needed--;
             pegasus_sleep(0);             pegasus_sleep(0);
          }          }
       }  
    }  
     return bodies;     return bodies;
 } }
  
Line 802 
Line 1022 
 { {
    if(th == 0)    if(th == 0)
       throw NullPointer();       throw NullPointer();
      try
      {
   
         timed_mutex(&(this->_monitor), 1000);
         if(_dying.value())
         {
            th->cancel();
            th->join();
            delete th;
         }
   
    _pool.insert_first(th);    _pool.insert_first(th);
   
      }
      catch(...)
      {
      }
 } }
  
  


Legend:
Removed from v.1.36.4.6  
changed lines
  Added in v.1.56

No CVS admin address has been configured
Powered by
ViewCVS 0.9.2