(file) Return to MessageQueueService.cpp CVS log (file) (dir) Up to [Pegasus] / pegasus / src / Pegasus / Common

Diff for /pegasus/src/Pegasus/Common/MessageQueueService.cpp between version 1.31 and 1.88.2.7

version 1.31, 2002/03/08 01:57:42 version 1.88.2.7, 2005/09/02 17:44:00
Line 1 
Line 1 
 //%////-*-c++-*-////////////////////////////////////////////////////////////////  //%2005////////////////////////////////////////////////////////////////////////
 // //
 // Copyright (c) 2000, 2001 The Open group, BMC Software, Tivoli Systems, IBM  // Copyright (c) 2000, 2001, 2002 BMC Software; Hewlett-Packard Development
   // Company, L.P.; IBM Corp.; The Open Group; Tivoli Systems.
   // Copyright (c) 2003 BMC Software; Hewlett-Packard Development Company, L.P.;
   // IBM Corp.; EMC Corporation, The Open Group.
   // Copyright (c) 2004 BMC Software; Hewlett-Packard Development Company, L.P.;
   // IBM Corp.; EMC Corporation; VERITAS Software Corporation; The Open Group.
   // Copyright (c) 2005 Hewlett-Packard Development Company, L.P.; IBM Corp.;
   // EMC Corporation; VERITAS Software Corporation; The Open Group.
 // //
 // Permission is hereby granted, free of charge, to any person obtaining a copy // Permission is hereby granted, free of charge, to any person obtaining a copy
 // of this software and associated documentation files (the "Software"), to // of this software and associated documentation files (the "Software"), to
Line 23 
Line 30 
 // Author: Mike Day (mdday@us.ibm.com) // Author: Mike Day (mdday@us.ibm.com)
 // //
 // Modified By: // Modified By:
   //              Amit K Arora, IBM (amita@in.ibm.com) for Bug#1090
 // //
 //%///////////////////////////////////////////////////////////////////////////// //%/////////////////////////////////////////////////////////////////////////////
  
 #include "MessageQueueService.h" #include "MessageQueueService.h"
 #include <Pegasus/Common/Tracer.h> #include <Pegasus/Common/Tracer.h>
   #include <Pegasus/Common/MessageLoader.h> //l10n
  
 PEGASUS_NAMESPACE_BEGIN PEGASUS_NAMESPACE_BEGIN
  
Line 35 
Line 44 
 cimom *MessageQueueService::_meta_dispatcher = 0; cimom *MessageQueueService::_meta_dispatcher = 0;
 AtomicInt MessageQueueService::_service_count = 0; AtomicInt MessageQueueService::_service_count = 0;
 AtomicInt MessageQueueService::_xid(1); AtomicInt MessageQueueService::_xid(1);
 Mutex MessageQueueService::_meta_dispatcher_mutex  = Mutex();  Mutex MessageQueueService::_meta_dispatcher_mutex;
   
   static struct timeval create_time = {0, 1};
   static struct timeval destroy_time = {300, 0};
   static struct timeval deadlock_time = {0, 0};
   
   ThreadPool *MessageQueueService::_thread_pool = 0;
   
   DQueue<MessageQueueService>* MessageQueueService::_polling_list = 0;
   
   Thread* MessageQueueService::_polling_thread = 0;
   
   ThreadPool *MessageQueueService::get_thread_pool(void)
   {
      return _thread_pool;
   }
   
   //
   // MAX_THREADS_PER_SVC_QUEUE_LIMIT
   //
   // JR Wunderlich Jun 6, 2005
   //
   
   #define MAX_THREADS_PER_SVC_QUEUE_LIMIT 5000
   #define MAX_THREADS_PER_SVC_QUEUE_DEFAULT 5
   
   Uint32 max_threads_per_svc_queue;
   
   PEGASUS_THREAD_RETURN PEGASUS_THREAD_CDECL  MessageQueueService::kill_idle_threads(void *parm)
   {
   
      static struct timeval now, last = {0,0};
      gettimeofday(&now, NULL);
      int dead_threads = 0;
   
      if( now.tv_sec - last.tv_sec > 120 )
      {
         gettimeofday(&last, NULL);
         try
         {
            dead_threads =  MessageQueueService::_thread_pool->kill_dead_threads();
         }
         catch(...)
         {
   
         }
      }
   
   #ifdef PEGASUS_POINTER_64BIT
      return (PEGASUS_THREAD_RETURN)(Uint64)dead_threads;
   #elif PEGASUS_PLATFORM_AIX_RS_IBMCXX
      return (PEGASUS_THREAD_RETURN)(unsigned long)dead_threads;
   #else
      return (PEGASUS_THREAD_RETURN)(Uint32)dead_threads;
   #endif
   }
   
   
   void MessageQueueService::force_shutdown(Boolean destroy_flag)
   {
      return;
   
   #ifdef MESSAGEQUEUESERVICE_DEBUG
           //l10n
      MessageLoaderParms parms("Common.MessageQueueService.FORCING_SHUTDOWN",
                               "Forcing shutdown of CIMOM Message Router");
      PEGASUS_STD(cout) << MessageLoader::getMessage(parms) << PEGASUS_STD(endl);
   #endif
   
   
      MessageQueueService *svc;
      int counter = 0;
      _polling_list->lock();
      svc = _polling_list->next(0);
   
      while(svc != 0)
      {
   #ifdef MESSAGEQUEUESERVICE_DEBUG
                   //l10n - reuse same MessageLoaderParms to avoid multiple creates
           parms.msg_id = "Common.MessageQueueService.STOPPING_SERVICE";
                   parms.default_msg = "Stopping $0";
                   parms.arg0 = svc->getQueueName();
                   PEGASUS_STD(cout) << MessageLoader::getMessage(parms) << PEGASUS_STD(endl);
   #endif
   
         _polling_sem->signal();
         svc->_shutdown_incoming_queue();
         counter++;
         _polling_sem->signal();
         svc = _polling_list->next(svc);
      }
      _polling_list->unlock();
   
      _polling_sem->signal();
   
      *MessageQueueService::_stop_polling = 1;
   
      if(destroy_flag == true)
      {
   
         svc = _polling_list->remove_last();
         while(svc)
         {
            delete svc;
            svc = _polling_list->remove_last();
         }
   
      }
   }
   
   
   PEGASUS_THREAD_RETURN PEGASUS_THREAD_CDECL MessageQueueService::polling_routine(void *parm)
   {
      Thread *myself = reinterpret_cast<Thread *>(parm);
      DQueue<MessageQueueService> *list = reinterpret_cast<DQueue<MessageQueueService> *>(myself->get_parm());
      while ( _stop_polling->value()  == 0 )
      {
         _polling_sem->wait();
         if(_stop_polling->value() != 0 )
         {
            break;
         }
   
         // The polling_routine thread must hold the lock on the
         // _polling_thread list while processing incoming messages.
         // This lock is used to give this thread ownership of
         // services on the _polling_routine list.
   
         // This is necessary to avoid confict with other threads
         // processing the _polling_list
         // (e.g., MessageQueueServer::~MessageQueueService).
   
         list->lock();
         MessageQueueService *service = list->next(0);
         ThreadStatus rtn = PEGASUS_THREAD_OK;
         while (service != NULL)
         {
             if ((service->_incoming.count() > 0) &&
                 (service->_die.value() == 0) &&
                 (service->_threads < max_threads_per_svc_queue))
             {
                // The _threads count is used to track the
                // number of active threads that have been allocated
                // to process messages for this service.
   
                // The _threads count MUST be incremented while
                // the polling_routine owns the _polling_thread
                // lock and has ownership of the service object.
   
                service->_threads++;
                try
                {
                    rtn = _thread_pool->allocate_and_awaken(
                         service, _req_proc, _polling_sem);
                }
                catch (...)
                {
                    service->_threads--;
   
                    // allocate_and_awaken should never generate an exception.
                    PEGASUS_ASSERT(0);
                }
                // if no more threads available, break from processing loop
                if (rtn != PEGASUS_THREAD_OK )
                {
                    service->_threads--;
                    Logger::put(Logger::STANDARD_LOG, System::CIMSERVER, Logger::TRACE,
                       "Not enough threads to process this request. Skipping.");
   
                    Tracer::trace(TRC_MESSAGEQUEUESERVICE, Tracer::LEVEL2,
                       "Could not allocate thread for %s. " \
                       "Queue has %d messages waiting and %d threads servicing." \
                       "Skipping the service for right now. ",
                       service->getQueueName(),
                       service->_incoming.count(),
                       service->_threads.value());
   
                    pegasus_yield();
                    service = NULL;
                 }
             }
             if (service != NULL)
             {
                service = list->next(service);
             }
         }
         list->unlock();
   
         if(_check_idle_flag->value() != 0 )
         {
            *_check_idle_flag = 0;
   
            // try to do idle thread clean up processing when system is not busy
            // if system is busy there may not be a thread available to allocate
            // so nothing will be done and that is OK.
   
            if ( _thread_pool->allocate_and_awaken(service, kill_idle_threads,
                 _polling_sem) != PEGASUS_THREAD_OK)
            {
                Logger::put(Logger::STANDARD_LOG, System::CIMSERVER, Logger::TRACE,
                   "Not enough threads to kill idle threads. What an irony.");
   
                Tracer::trace(TRC_MESSAGEQUEUESERVICE, Tracer::LEVEL2,
                   "Could not allocate thread to kill idle threads." \
                   "Skipping. ");
            }
         }
      }
      myself->exit_self( (PEGASUS_THREAD_RETURN) 1 );
      return(0);
   }
   
   
   Semaphore* MessageQueueService::_polling_sem = 0;
   AtomicInt* MessageQueueService::_stop_polling = 0;
   AtomicInt* MessageQueueService::_check_idle_flag = 0;
  
  
 MessageQueueService::MessageQueueService(const char *name, MessageQueueService::MessageQueueService(const char *name,
Line 46 
Line 270 
  
      _mask(mask),      _mask(mask),
      _die(0),      _die(0),
      _pending(true),       _threads(0),
      _incoming(true, 1000),       _incoming(true, 0),
        _callback(true),
      _incoming_queue_shutdown(0),      _incoming_queue_shutdown(0),
      _req_thread(_req_proc, this, false)       _callback_ready(0),
        _req_thread(_req_proc, this, false),
        _callback_thread(_callback_proc, this, false)
   
 { {
   
    _capabilities = (capabilities | module_capabilities::async);    _capabilities = (capabilities | module_capabilities::async);
  
    _default_op_timeout.tv_sec = 30;    _default_op_timeout.tv_sec = 30;
    _default_op_timeout.tv_usec = 100;    _default_op_timeout.tv_usec = 100;
  
    _meta_dispatcher_mutex.lock(pegasus_thread_self());     max_threads_per_svc_queue = MAX_THREADS_PER_SVC_QUEUE;
   
      // if requested threads gt MAX_THREADS_PER_SVC_QUEUE_LIMIT
      // then set to MAX_THREADS_PER_SVC_QUEUE_LIMIT
   
      if (max_threads_per_svc_queue > MAX_THREADS_PER_SVC_QUEUE_LIMIT)
      {
         max_threads_per_svc_queue = MAX_THREADS_PER_SVC_QUEUE_LIMIT;
      }
   
      // if requested threads eq 0 (unlimited)
      // then set to MAX_THREADS_PER_SVC_QUEUE_LIMIT
   
      if (max_threads_per_svc_queue == 0)
      {
         max_threads_per_svc_queue = MAX_THREADS_PER_SVC_QUEUE_DEFAULT;
      }
   
      // cout << "MAX_THREADS_PER_SVC_QUEUE = " << MAX_THREADS_PER_SVC_QUEUE << endl;
      // cout << "max_threads_per_svc_queue set to = " << max_threads_per_svc_queue << endl;
   
      AutoMutex autoMut(_meta_dispatcher_mutex);
  
    if( _meta_dispatcher == 0 )    if( _meta_dispatcher == 0 )
    {    {
         // Instantiate the common objects
         _polling_list = new DQueue<MessageQueueService>(true);
         _stop_polling = new AtomicInt(0);
         _polling_sem = new Semaphore(0);
         _check_idle_flag = new AtomicInt(0);
   
         *_stop_polling = 0;
       PEGASUS_ASSERT( _service_count.value() == 0 );       PEGASUS_ASSERT( _service_count.value() == 0 );
       _meta_dispatcher = new cimom();       _meta_dispatcher = new cimom();
       if (_meta_dispatcher == NULL )       if (_meta_dispatcher == NULL )
       {       {
          _meta_dispatcher_mutex.unlock();  
   
          throw NullPointer();          throw NullPointer();
       }       }
         _thread_pool = new ThreadPool(0, "MessageQueueService", 0, 0,
                                       create_time, destroy_time, deadlock_time);
  
         _polling_thread = new Thread(polling_routine,
                                      reinterpret_cast<void *>(_polling_list),
                                      false);
         while (!_polling_thread->run())
         {
            pegasus_yield();
         }
    }    }
    _service_count++;    _service_count++;
  
   
    if( false == register_service(name, _capabilities, _mask) )    if( false == register_service(name, _capabilities, _mask) )
    {    {
       _meta_dispatcher_mutex.unlock();        //l10n
       throw BindFailed("MessageQueueService Base Unable to register with  Meta Dispatcher");        //throw BindFailedException("MessageQueueService Base Unable to register with  Meta Dispatcher");
         MessageLoaderParms parms("Common.MessageQueueService.UNABLE_TO_REGISTER",
                                  "MessageQueueService Base Unable to register with  Meta Dispatcher");
   
         throw BindFailedException(parms);
    }    }
  
    _meta_dispatcher_mutex.unlock();     _polling_list->insert_last(this);
  
    _req_thread.run();  //   _meta_dispatcher_mutex.unlock();  //Bug#1090
   //   _callback_thread.run();
   
   //   _req_thread.run();
 } }
  
  
 MessageQueueService::~MessageQueueService(void) MessageQueueService::~MessageQueueService(void)
 { {
    _die = 1;    _die = 1;
   
      // The polling_routine locks the _polling_list while
      // processing the incoming messages for services on the
      // list.  Deleting the service from the _polling_list
      // prior to processing, avoids synchronization issues
      // with the _polling_routine.
   
      _polling_list->remove(this);
   
      _callback_ready.signal();
   
      // ATTN: The code for closing the _incoming queue
      // is not working correctly. In OpenPegasus 2.4,
      // execution of the following code is very timing
      // dependent. This needs to be fix.
      // See Bug 4079 for details.
    if (_incoming_queue_shutdown.value() == 0 )    if (_incoming_queue_shutdown.value() == 0 )
    {    {
        _incoming.shutdown_queue();         _shutdown_incoming_queue();
        _req_thread.join();  
    }    }
  
    _meta_dispatcher_mutex.lock(pegasus_thread_self());     // Wait until all threads processing the messages
      // for this service have completed.
   
      while (_threads.value() > 0)
      {
         pegasus_yield();
      }
   
      {
        AutoMutex autoMut(_meta_dispatcher_mutex);
    _service_count--;    _service_count--;
    if (_service_count.value() == 0 )    if (_service_count.value() == 0 )
    {    {
   
         (*_stop_polling)++;
         _polling_sem->signal();
         _polling_thread->join();
         delete _polling_thread;
         _polling_thread = 0;
       _meta_dispatcher->_shutdown_routed_queue();       _meta_dispatcher->_shutdown_routed_queue();
       delete _meta_dispatcher;       delete _meta_dispatcher;
    }        _meta_dispatcher = 0;
    _meta_dispatcher_mutex.unlock();  
   
 }  
  
         delete _thread_pool;
         _thread_pool = 0;
  
         // Clean up the common objects
         delete _check_idle_flag;
         delete _polling_sem;
         delete _stop_polling;
         delete _polling_list;
        }
      } // mutex unlocks here
      // Clean up in case there are extra stuff on the queue.
      while (_incoming.count())
      {
          try
          {
             delete _incoming.remove_first();
          }
          catch (const ListClosed &e)
          {
             // If the list is closed, there is nothing we can do.
             break;
          }
      }
   }
  
 void MessageQueueService::_shutdown_incoming_queue(void) void MessageQueueService::_shutdown_incoming_queue(void)
 { {
  
   
    if (_incoming_queue_shutdown.value() > 0 )    if (_incoming_queue_shutdown.value() > 0 )
       return ;       return ;
   
    AsyncIoctl *msg = new AsyncIoctl(get_next_xid(),    AsyncIoctl *msg = new AsyncIoctl(get_next_xid(),
                                     0,                                     0,
                                     _queueId,                                     _queueId,
Line 123 
Line 443 
                                     0);                                     0);
  
    msg->op = get_op();    msg->op = get_op();
    msg->op->_request.insert_first(msg);     msg->op->_flags |= ASYNC_OPFLAGS_FIRE_AND_FORGET;
      msg->op->_flags &= ~(ASYNC_OPFLAGS_CALLBACK | ASYNC_OPFLAGS_SAFE_CALLBACK
                      | ASYNC_OPFLAGS_SIMPLE_STATUS);
      msg->op->_state &= ~ASYNC_OPSTATE_COMPLETE;
   
    msg->op->_op_dest = this;    msg->op->_op_dest = this;
      msg->op->_request.insert_first(msg);
  
      try
      {
    _incoming.insert_last_wait(msg->op);    _incoming.insert_last_wait(msg->op);
    msg->op->_client_sem.wait();         _polling_sem->signal();
      }
    msg->op->lock();     catch (const ListClosed &)
    AsyncReply * reply = static_cast<AsyncReply *>(msg->op->_response.remove_first());     {
    reply->op = 0;         // This means the queue has already been shut-down (happens  when there
    msg->op->unlock();         // are two AsyncIoctrl::IO_CLOSE messages generated and one got first
    delete reply;         // processed.
   
    msg->op->_request.remove(msg);  
    msg->op->_state |= ASYNC_OPSTATE_RELEASED;  
    return_op(msg->op);  
   
    msg->op = 0;  
    delete msg;    delete msg;
    _req_thread.join();     }
      catch (const Permission &)
      {
          delete msg;
      }
 } }
  
  
  
 void MessageQueueService::enqueue(Message *msg) throw(IPCException)  void MessageQueueService::enqueue(Message *msg)
 { {
    PEG_METHOD_ENTER(TRC_MESSAGEQUEUESERVICE, "MessageQueueService::enqueue()");    PEG_METHOD_ENTER(TRC_MESSAGEQUEUESERVICE, "MessageQueueService::enqueue()");
  
Line 157 
Line 481 
 } }
  
  
   PEGASUS_THREAD_RETURN PEGASUS_THREAD_CDECL MessageQueueService::_callback_proc(void *parm)
 PEGASUS_THREAD_RETURN PEGASUS_THREAD_CDECL MessageQueueService::_req_proc(void * parm)  
 { {
    Thread *myself = reinterpret_cast<Thread *>(parm);    Thread *myself = reinterpret_cast<Thread *>(parm);
    MessageQueueService *service = reinterpret_cast<MessageQueueService *>(myself->get_parm());    MessageQueueService *service = reinterpret_cast<MessageQueueService *>(myself->get_parm());
      AsyncOpNode *operation = 0;
  
      while ( service->_die.value() == 0 )
      {
         service->_callback_ready.wait();
   
         service->_callback.lock();
         operation = service->_callback.next(0);
         while( operation != NULL)
         {
            if( ASYNC_OPSTATE_COMPLETE & operation->read_state())
            {
               operation = service->_callback.remove_no_lock(operation);
               PEGASUS_ASSERT(operation != NULL);
               operation->_thread_ptr = myself;
               operation->_service_ptr = service;
               service->_handle_async_callback(operation);
               break;
            }
            operation = service->_callback.next(operation);
         }
         service->_callback.unlock();
      }
      myself->exit_self( (PEGASUS_THREAD_RETURN) 1 );
      return(0);
   }
   
   
   PEGASUS_THREAD_RETURN PEGASUS_THREAD_CDECL MessageQueueService::_req_proc(void * parm)
   {
      MessageQueueService *service = reinterpret_cast<MessageQueueService *>(parm);
    // pull messages off the incoming queue and dispatch them. then    // pull messages off the incoming queue and dispatch them. then
    // check pending messages that are non-blocking    // check pending messages that are non-blocking
    AsyncOpNode *operation = 0;    AsyncOpNode *operation = 0;
  
    while ( service->_die.value() == 0 )     if ( service->_die.value() == 0 )
    {    {
       try       try
       {       {
          operation = service->_incoming.remove_first_wait();              operation = service->_incoming.remove_first();
       }       }
       catch(ListClosed & )       catch(ListClosed & )
       {       {
          break;              operation = 0;
               service->_threads--;
               return(0);
       }       }
       if( operation )       if( operation )
       {       {
               operation->_service_ptr = service;
          service->_handle_incoming_operation(operation, myself, service);              service->_handle_incoming_operation(operation);
       }       }
    }    }
      service->_threads--;
    myself->exit_self( (PEGASUS_THREAD_RETURN) 1 );  
    return(0);    return(0);
 } }
  
   Uint32 MessageQueueService::get_pending_callback_count(void)
   {
      return _callback.count();
   }
   
   
   
   void MessageQueueService::_sendwait_callback(AsyncOpNode *op,
                                                MessageQueue *q,
                                                void *parm)
   {
      op->_client_sem.signal();
   }
   
  
 // callback function is responsible for cleaning up all resources // callback function is responsible for cleaning up all resources
 // including op, op->_callback_node, and op->_callback_ptr // including op, op->_callback_node, and op->_callback_ptr
 void MessageQueueService::_handle_async_callback(AsyncOpNode *op) void MessageQueueService::_handle_async_callback(AsyncOpNode *op)
 { {
      if( op->_flags & ASYNC_OPFLAGS_SAFE_CALLBACK )
      {
   
         Message *msg = op->get_request();
         if( msg && ( msg->getMask() & message_mask::ha_async))
         {
            if(msg->getType() == async_messages::ASYNC_LEGACY_OP_START )
            {
               AsyncLegacyOperationStart *wrapper =
                  static_cast<AsyncLegacyOperationStart *>(msg);
               msg = wrapper->get_action();
               delete wrapper;
            }
            else if (msg->getType() == async_messages::ASYNC_MODULE_OP_START)
            {
               AsyncModuleOperationStart *wrapper =
                  static_cast<AsyncModuleOperationStart *>(msg);
               msg = wrapper->get_action();
               delete wrapper;
            }
            else if (msg->getType() == async_messages::ASYNC_OP_START)
            {
               AsyncOperationStart *wrapper =
                  static_cast<AsyncOperationStart *>(msg);
               msg = wrapper->get_action();
               delete wrapper;
            }
            delete msg;
         }
   
         msg = op->get_response();
         if( msg && ( msg->getMask() & message_mask::ha_async))
         {
            if(msg->getType() == async_messages::ASYNC_LEGACY_OP_RESULT )
            {
               AsyncLegacyOperationResult *wrapper =
                  static_cast<AsyncLegacyOperationResult *>(msg);
               msg = wrapper->get_result();
               delete wrapper;
            }
            else if (msg->getType() == async_messages::ASYNC_MODULE_OP_RESULT)
            {
               AsyncModuleOperationResult *wrapper =
                  static_cast<AsyncModuleOperationResult *>(msg);
               msg = wrapper->get_result();
               delete wrapper;
            }
         }
         void (*callback)(Message *, void *, void *) = op->__async_callback;
         void *handle = op->_callback_handle;
         void *parm = op->_callback_parameter;
         op->release();
         return_op(op);
         callback(msg, handle, parm);
      }
      else if( op->_flags & ASYNC_OPFLAGS_CALLBACK )
      {
    // note that _callback_node may be different from op    // note that _callback_node may be different from op
    // op->_callback_q is a "this" pointer we can use for static callback methods        // op->_callback_response_q is a "this" pointer we can use for
    op->_async_callback(op->_callback_node, op->_callback_q, op->_callback_ptr);        // static callback methods
         op->_async_callback(op->_callback_node, op->_callback_response_q, op->_callback_ptr);
      }
 } }
  
  
 void MessageQueueService::_handle_incoming_operation(AsyncOpNode *operation,  void MessageQueueService::_handle_incoming_operation(AsyncOpNode *operation)
                                                      Thread *thread,  //                                                   Thread *thread,
                                                      MessageQueue *queue)  //                                                   MessageQueue *queue)
 { {
    if ( operation != 0 )    if ( operation != 0 )
    {    {
Line 222 
Line 649 
          rq = operation->_request.remove_first() ;          rq = operation->_request.remove_first() ;
          operation->unlock();          operation->unlock();
          // delete the op node          // delete the op node
          delete operation;           operation->release();
            return_op( operation);
  
          handleEnqueue(rq);          handleEnqueue(rq);
          return;          return;
       }       }
  
       if ( operation->_state & ASYNC_OPFLAGS_CALLBACK &&        if ( (operation->_flags & ASYNC_OPFLAGS_CALLBACK ||
               operation->_flags & ASYNC_OPFLAGS_SAFE_CALLBACK) &&
            (operation->_state & ASYNC_OPSTATE_COMPLETE))            (operation->_state & ASYNC_OPSTATE_COMPLETE))
       {       {
   
          operation->unlock();          operation->unlock();
          _handle_async_callback(operation);          _handle_async_callback(operation);
       }       }
       else       else
       {       {
          PEGASUS_ASSERT(rq != 0 );          PEGASUS_ASSERT(rq != 0 );
          // ATTN: optimization  
          // << Wed Mar  6 15:00:39 2002 mdd >>  
          // put thread and queue into the asyncopnode structure.  
          (static_cast<AsyncMessage *>(rq))->_myself = thread;  
          (static_cast<AsyncMessage *>(rq))->_service = queue;  
          operation->unlock();          operation->unlock();
          _handle_async_request(static_cast<AsyncRequest *>(rq));          _handle_async_request(static_cast<AsyncRequest *>(rq));
       }       }
Line 284 
Line 709 
    Message* response)    Message* response)
  
 { {
      PEG_METHOD_ENTER(TRC_MESSAGEQUEUESERVICE,
                       "MessageQueueService::_enqueueResponse");
  
    if( request->getMask() & message_mask::ha_async)    if( request->getMask() & message_mask::ha_async)
    {    {
Line 292 
Line 719 
          _completeAsyncResponse(static_cast<AsyncRequest *>(request),          _completeAsyncResponse(static_cast<AsyncRequest *>(request),
                                 static_cast<AsyncReply *>(response),                                 static_cast<AsyncReply *>(response),
                                 ASYNC_OPSTATE_COMPLETE, 0 );                                 ASYNC_OPSTATE_COMPLETE, 0 );
            PEG_METHOD_EXIT();
          return true;          return true;
       }       }
    }    }
Line 304 
Line 732 
       AsyncRequest *async = static_cast<AsyncRequest *>(request->_async);       AsyncRequest *async = static_cast<AsyncRequest *>(request->_async);
       AsyncOpNode *op = async->op;       AsyncOpNode *op = async->op;
       request->_async = 0;       request->_async = 0;
       // this request is probably going to be deleted !!        // the legacy request is going to be deleted by its handler
       // remove it from the op node       // remove it from the op node
       op->_request.remove(request);  
         static_cast<AsyncLegacyOperationStart *>(async)->get_action();
   
  
       AsyncLegacyOperationResult *async_result =       AsyncLegacyOperationResult *async_result =
          new AsyncLegacyOperationResult(          new AsyncLegacyOperationResult(
Line 318 
Line 748 
                              async_result,                              async_result,
                              ASYNC_OPSTATE_COMPLETE,                              ASYNC_OPSTATE_COMPLETE,
                              0);                              0);
         PEG_METHOD_EXIT();
       return true;       return true;
    }    }
  
    // ensure that the destination queue is in response->dest    // ensure that the destination queue is in response->dest
      PEG_METHOD_EXIT();
    return SendForget(response);    return SendForget(response);
  
 } }
Line 337 
Line 769 
                                                 Uint32 state,                                                 Uint32 state,
                                                 Uint32 flag)                                                 Uint32 flag)
 { {
      PEG_METHOD_ENTER(TRC_MESSAGEQUEUESERVICE,
                       "MessageQueueService::_completeAsyncResponse");
   
    cimom::_completeAsyncResponse(request, reply, state, flag);    cimom::_completeAsyncResponse(request, reply, state, flag);
   
      PEG_METHOD_EXIT();
 } }
  
  
   void MessageQueueService::_complete_op_node(AsyncOpNode *op,
                                               Uint32 state,
                                               Uint32 flag,
                                               Uint32 code)
   {
      cimom::_complete_op_node(op, state, flag, code);
   }
   
  
 Boolean MessageQueueService::accept_async(AsyncOpNode *op) Boolean MessageQueueService::accept_async(AsyncOpNode *op)
 { {
    if (_incoming_queue_shutdown.value() > 0 )    if (_incoming_queue_shutdown.value() > 0 )
       return false;        return true;
  
 // ATTN optimization remove the message checking altogether in the base // ATTN optimization remove the message checking altogether in the base
 // << Mon Feb 18 14:02:20 2002 mdd >> // << Mon Feb 18 14:02:20 2002 mdd >>
Line 358 
Line 803 
         _die.value() == 0  )         _die.value() == 0  )
    {    {
       _incoming.insert_last_wait(op);       _incoming.insert_last_wait(op);
         _polling_sem->signal();
       return true;       return true;
    }    }
    return false;    return false;
Line 370 
Line 816 
    return true;    return true;
 } }
  
   
 // made pure virtual  
 // << Wed Mar  6 15:11:31 2002 mdd >>  
 // void MessageQueueService::handleEnqueue(Message *msg)  
 // {  
 //    if ( msg )  
 //       delete msg;  
 // }  
   
 // made pure virtual  
 // << Wed Mar  6 15:11:56 2002 mdd >>  
 // void MessageQueueService::handleEnqueue(void)  
 // {  
 //     Message *msg = dequeue();  
 //     handleEnqueue(msg);  
 // }  
   
 void MessageQueueService::handle_heartbeat_request(AsyncRequest *req) void MessageQueueService::handle_heartbeat_request(AsyncRequest *req)
 { {
    // default action is to echo a heartbeat response    // default action is to echo a heartbeat response
Line 416 
Line 845 
    {    {
       case AsyncIoctl::IO_CLOSE:       case AsyncIoctl::IO_CLOSE:
       {       {
          // save my bearings  
          Thread *myself = req->_myself;  
          MessageQueueService *service = static_cast<MessageQueueService *>(req->_service);  
  
          // respond to this message.           MessageQueueService *service = static_cast<MessageQueueService *>(req->op->_service_ptr);
   
   #ifdef MESSAGEQUEUESERVICE_DEBUG
            PEGASUS_STD(cout) << service->getQueueName() << " Received AsyncIoctl::IO_CLOSE " << PEGASUS_STD(endl);
   #endif
   
            // respond to this message. this is fire and forget, so we don't need to delete anything.
            // this takes care of two problems that were being found
            // << Thu Oct  9 10:52:48 2003 mdd >>
          _make_response(req, async_results::OK);          _make_response(req, async_results::OK);
          // ensure we do not accept any further messages          // ensure we do not accept any further messages
  
Line 444 
Line 878 
             }             }
             if( operation )             if( operation )
             {             {
                service->_handle_incoming_operation(operation, myself, service);                 operation->_service_ptr = service;
                  service->_handle_incoming_operation(operation);
             }             }
             else             else
                break;                break;
Line 452 
Line 887 
  
          // shutdown the AsyncDQueue          // shutdown the AsyncDQueue
          service->_incoming.shutdown_queue();          service->_incoming.shutdown_queue();
          // exit the thread !  
          myself->exit_self( (PEGASUS_THREAD_RETURN) 1 );  
          return;          return;
       }       }
  
Line 464 
Line 897 
  
 void MessageQueueService::handle_CimServiceStart(CimServiceStart *req) void MessageQueueService::handle_CimServiceStart(CimServiceStart *req)
 { {
   
   #ifdef MESSAGEQUEUESERVICE_DEBUG
      PEGASUS_STD(cout) << getQueueName() << "received START" << PEGASUS_STD(endl);
   #endif
   
    // clear the stoped bit and update    // clear the stoped bit and update
    _capabilities &= (~(module_capabilities::stopped));    _capabilities &= (~(module_capabilities::stopped));
    _make_response(req, async_results::OK);    _make_response(req, async_results::OK);
Line 473 
Line 911 
 } }
 void MessageQueueService::handle_CimServiceStop(CimServiceStop *req) void MessageQueueService::handle_CimServiceStop(CimServiceStop *req)
 { {
   #ifdef MESSAGEQUEUESERVICE_DEBUG
      PEGASUS_STD(cout) << getQueueName() << "received STOP" << PEGASUS_STD(endl);
   #endif
    // set the stopeed bit and update    // set the stopeed bit and update
    _capabilities |= module_capabilities::stopped;    _capabilities |= module_capabilities::stopped;
    _make_response(req, async_results::CIM_STOPPED);    _make_response(req, async_results::CIM_STOPPED);
Line 586 
Line 1027 
    // get the queue handle for the destination    // get the queue handle for the destination
  
    op->lock();    op->lock();
    op->_op_dest = MessageQueue::lookup(destination);     op->_op_dest = MessageQueue::lookup(destination); // destination of this message
    op->_flags |= ASYNC_OPFLAGS_CALLBACK;    op->_flags |= ASYNC_OPFLAGS_CALLBACK;
    op->_flags &= ~(ASYNC_OPFLAGS_FIRE_AND_FORGET);    op->_flags &= ~(ASYNC_OPFLAGS_FIRE_AND_FORGET);
    op->_state &= ~ASYNC_OPSTATE_COMPLETE;    op->_state &= ~ASYNC_OPSTATE_COMPLETE;
    // initialize the callback data    // initialize the callback data
    op->_async_callback = callback;     op->_async_callback = callback;   // callback function to be executed by recpt. of response
    op->_callback_node = op;     op->_callback_node = op;          // the op node
    op->_callback_response_q = callback_response_q;     op->_callback_response_q = callback_response_q;  // the queue that will receive the response
    op->_callback_ptr = callback_ptr;     op->_callback_ptr = callback_ptr;   // user data for callback
    op->_callback_q = this;     op->_callback_request_q = this;     // I am the originator of this request
  
    op->unlock();    op->unlock();
    if(op->_op_dest == 0)    if(op->_op_dest == 0)
Line 605 
Line 1046 
 } }
  
  
   Boolean MessageQueueService::SendAsync(Message *msg,
                                          Uint32 destination,
                                          void (*callback)(Message *response,
                                                           void *handle,
                                                           void *parameter),
                                          void *handle,
                                          void *parameter)
   {
      if(msg == NULL)
         return false;
      if(callback == NULL)
         return SendForget(msg);
      AsyncOpNode *op = get_op();
      msg->dest = destination;
      if( NULL == (op->_op_dest = MessageQueue::lookup(msg->dest)))
      {
         op->release();
         return_op(op);
         return false;
      }
      op->_flags |= ASYNC_OPFLAGS_SAFE_CALLBACK;
      op->_flags &= ~(ASYNC_OPFLAGS_FIRE_AND_FORGET);
      op->_state &= ~ASYNC_OPSTATE_COMPLETE;
      op->__async_callback = callback;
      op->_callback_node = op;
      op->_callback_handle = handle;
      op->_callback_parameter = parameter;
      op->_callback_response_q = this;
   
   
      if( ! (msg->getMask() & message_mask::ha_async) )
      {
         AsyncLegacyOperationStart *wrapper =
            new AsyncLegacyOperationStart(get_next_xid(),
                                          op,
                                          destination,
                                          msg,
                                          destination);
      }
      else
      {
         op->_request.insert_first(msg);
         (static_cast<AsyncMessage *>(msg))->op = op;
      }
   
      _callback.insert_last(op);
      return _meta_dispatcher->route_async(op);
   }
   
   
 Boolean MessageQueueService::SendForget(Message *msg) Boolean MessageQueueService::SendForget(Message *msg)
 { {
  
Line 624 
Line 1115 
       if (mask & message_mask::ha_async)       if (mask & message_mask::ha_async)
          (static_cast<AsyncMessage *>(msg))->op = op;          (static_cast<AsyncMessage *>(msg))->op = op;
    }    }
    op->lock();  
    op->_op_dest = MessageQueue::lookup(msg->dest);    op->_op_dest = MessageQueue::lookup(msg->dest);
    op->_flags |= ASYNC_OPFLAGS_FIRE_AND_FORGET;    op->_flags |= ASYNC_OPFLAGS_FIRE_AND_FORGET;
    op->_flags &= ~(ASYNC_OPFLAGS_CALLBACK | ASYNC_OPFLAGS_SIMPLE_STATUS);     op->_flags &= ~(ASYNC_OPFLAGS_CALLBACK | ASYNC_OPFLAGS_SAFE_CALLBACK
                      | ASYNC_OPFLAGS_SIMPLE_STATUS);
    op->_state &= ~ASYNC_OPSTATE_COMPLETE;    op->_state &= ~ASYNC_OPSTATE_COMPLETE;
    op->unlock();  
    if ( op->_op_dest == 0 )    if ( op->_op_dest == 0 )
      {
         op->release();
         return_op(op);
       return false;       return false;
      }
  
    // now see if the meta dispatcher will take it    // now see if the meta dispatcher will take it
    return  _meta_dispatcher->route_async(op);    return  _meta_dispatcher->route_async(op);
Line 645 
Line 1139 
  
    Boolean destroy_op = false;    Boolean destroy_op = false;
  
    if (request->op == false)     if (request->op == 0)
    {    {
       request->op = get_op();       request->op = get_op();
       request->op->_request.insert_first(request);       request->op->_request.insert_first(request);
       destroy_op = true;       destroy_op = true;
    }    }
  
    request->block = true;     request->block = false;
    request->op->lock();     request->op->_flags |= ASYNC_OPFLAGS_PSEUDO_CALLBACK;
    request->op->_state &= ~ASYNC_OPSTATE_COMPLETE;     SendAsync(request->op,
    request->op->_flags &= ~ASYNC_OPFLAGS_CALLBACK;               request->dest,
                _sendwait_callback,
    request->op->_op_dest = MessageQueue::lookup(request->dest);               this,
    request->op->unlock();               (void *)0);
   
    if ( request->op->_op_dest == 0 )  
       return 0;  
  
    // now see if the meta dispatcher will take it  
    if (true == _meta_dispatcher->route_async(request->op))  
    {  
       request->op->_client_sem.wait();       request->op->_client_sem.wait();
       PEGASUS_ASSERT(request->op->_state & ASYNC_OPSTATE_COMPLETE);  
    }  
   
    request->op->lock();    request->op->lock();
    AsyncReply * rpl = static_cast<AsyncReply *>(request->op->_response.remove_first());    AsyncReply * rpl = static_cast<AsyncReply *>(request->op->_response.remove_first());
    rpl->op = 0;    rpl->op = 0;
Line 684 
Line 1169 
       return_op(request->op);       return_op(request->op);
       request->op = 0;       request->op = 0;
    }    }
   
    return rpl;    return rpl;
 } }
  
Line 701 
Line 1185 
                                                     capabilities,                                                     capabilities,
                                                     mask,                                                     mask,
                                                     _queueId);                                                     _queueId);
      msg->dest = CIMOM_Q_ID;
   
    Boolean registered = false;    Boolean registered = false;
    AsyncReply *reply = static_cast<AsyncReply *>(SendWait( msg ));    AsyncReply *reply = static_cast<AsyncReply *>(SendWait( msg ));
  
Line 780 
Line 1266 
                            capabilities,                            capabilities,
                            mask);                            mask);
  
      req->dest = CIMOM_Q_ID;
   
    AsyncMessage *reply = SendWait(req);    AsyncMessage *reply = SendWait(req);
    if(reply)    if(reply)
    {    {
Line 848 
Line 1336 
  
 Uint32 MessageQueueService::get_next_xid(void) Uint32 MessageQueueService::get_next_xid(void)
 { {
      static Mutex _monitor;
      Uint32 value;
      AutoMutex autoMut(_monitor);
    _xid++;    _xid++;
    return _xid.value();     value =  _xid.value();
      return value;
   
 } }
  
 PEGASUS_NAMESPACE_END PEGASUS_NAMESPACE_END


Legend:
Removed from v.1.31  
changed lines
  Added in v.1.88.2.7

No CVS admin address has been configured
Powered by
ViewCVS 0.9.2