(file) Return to MessageQueueService.cpp CVS log (file) (dir) Up to [Pegasus] / pegasus / src / Pegasus / Common

Diff for /pegasus/src/Pegasus/Common/MessageQueueService.cpp between version 1.114 and 1.131

version 1.114, 2005/07/01 20:15:59 version 1.131, 2006/11/10 18:14:58
Line 1 
Line 1 
 //%2005////////////////////////////////////////////////////////////////////////  //%2006////////////////////////////////////////////////////////////////////////
 // //
 // Copyright (c) 2000, 2001, 2002 BMC Software; Hewlett-Packard Development // Copyright (c) 2000, 2001, 2002 BMC Software; Hewlett-Packard Development
 // Company, L.P.; IBM Corp.; The Open Group; Tivoli Systems. // Company, L.P.; IBM Corp.; The Open Group; Tivoli Systems.
Line 8 
Line 8 
 // IBM Corp.; EMC Corporation; VERITAS Software Corporation; The Open Group. // IBM Corp.; EMC Corporation; VERITAS Software Corporation; The Open Group.
 // Copyright (c) 2005 Hewlett-Packard Development Company, L.P.; IBM Corp.; // Copyright (c) 2005 Hewlett-Packard Development Company, L.P.; IBM Corp.;
 // EMC Corporation; VERITAS Software Corporation; The Open Group. // EMC Corporation; VERITAS Software Corporation; The Open Group.
   // Copyright (c) 2006 Hewlett-Packard Development Company, L.P.; IBM Corp.;
   // EMC Corporation; Symantec Corporation; The Open Group.
 // //
 // Permission is hereby granted, free of charge, to any person obtaining a copy // Permission is hereby granted, free of charge, to any person obtaining a copy
 // of this software and associated documentation files (the "Software"), to // of this software and associated documentation files (the "Software"), to
Line 27 
Line 29 
 // //
 //============================================================================== //==============================================================================
 // //
 // Author: Mike Day (mdday@us.ibm.com)  
 //  
 // Modified By:  
 //              Amit K Arora, IBM (amita@in.ibm.com) for Bug#1090,#2657  
 //              Josephine Eskaline Joyce, IBM (jojustin@in.ibm.com) for Bug#3259  
 //              Jim Wunderlich (Jim_Wunderlich@prodigy.net)  
 //  
 //%///////////////////////////////////////////////////////////////////////////// //%/////////////////////////////////////////////////////////////////////////////
  
 // #include <iostream.h>  
 #include "MessageQueueService.h" #include "MessageQueueService.h"
 #include <Pegasus/Common/Tracer.h> #include <Pegasus/Common/Tracer.h>
 #include <Pegasus/Common/MessageLoader.h> //l10n  #include <Pegasus/Common/MessageLoader.h>
  
 PEGASUS_NAMESPACE_BEGIN PEGASUS_NAMESPACE_BEGIN
  
 cimom *MessageQueueService::_meta_dispatcher = 0; cimom *MessageQueueService::_meta_dispatcher = 0;
 AtomicInt MessageQueueService::_service_count = 0;  AtomicInt MessageQueueService::_service_count(0);
 AtomicInt MessageQueueService::_xid(1);  
 Mutex MessageQueueService::_meta_dispatcher_mutex; Mutex MessageQueueService::_meta_dispatcher_mutex;
  
 static struct timeval deallocateWait = {300, 0}; static struct timeval deallocateWait = {300, 0};
  
 ThreadPool *MessageQueueService::_thread_pool = 0; ThreadPool *MessageQueueService::_thread_pool = 0;
  
 DQueue<MessageQueueService> MessageQueueService::_polling_list(true);  MessageQueueService::PollingList* MessageQueueService::_polling_list;
   Mutex MessageQueueService::_polling_list_mutex;
  
 Thread* MessageQueueService::_polling_thread = 0; Thread* MessageQueueService::_polling_thread = 0;
  
Line 60 
Line 54 
 { {
    return _thread_pool;    return _thread_pool;
 } }
   
 // //
 // MAX_THREADS_PER_SVC_QUEUE_LIMIT  // MAX_THREADS_PER_SVC_QUEUE
 // //
 // JR Wunderlich Jun 6, 2005 // JR Wunderlich Jun 6, 2005
 // //
Line 69 
Line 64 
 #define MAX_THREADS_PER_SVC_QUEUE_LIMIT 5000 #define MAX_THREADS_PER_SVC_QUEUE_LIMIT 5000
 #define MAX_THREADS_PER_SVC_QUEUE_DEFAULT 5 #define MAX_THREADS_PER_SVC_QUEUE_DEFAULT 5
  
 Uint32 max_threads_per_svc_queue;  #ifndef MAX_THREADS_PER_SVC_QUEUE
   # define MAX_THREADS_PER_SVC_QUEUE MAX_THREADS_PER_SVC_QUEUE_DEFAULT
 PEGASUS_THREAD_RETURN PEGASUS_THREAD_CDECL  
 MessageQueueService::kill_idle_threads(void *parm)  
 {  
   
    static struct timeval now, last = {0,0};  
    gettimeofday(&now, NULL);  
    int dead_threads = 0;  
   
    if (now.tv_sec - last.tv_sec > 120)  
    {  
       gettimeofday(&last, NULL);  
       try  
       {  
          dead_threads = MessageQueueService::_thread_pool->cleanupIdleThreads();  
       }  
       catch(...)  
       {  
   
       }  
    }  
   
 #ifdef PEGASUS_POINTER_64BIT  
    return (PEGASUS_THREAD_RETURN)(Uint64)dead_threads;  
 #elif PEGASUS_PLATFORM_AIX_RS_IBMCXX  
    return (PEGASUS_THREAD_RETURN)(unsigned long)dead_threads;  
 #else  
    return (PEGASUS_THREAD_RETURN)(Uint32)dead_threads;  
 #endif #endif
 }  
  
 PEGASUS_THREAD_RETURN PEGASUS_THREAD_CDECL MessageQueueService::polling_routine(void *parm)  Uint32 max_threads_per_svc_queue;
   
   ThreadReturnType PEGASUS_THREAD_CDECL MessageQueueService::polling_routine(
       void* parm)
 { {
    Thread *myself = reinterpret_cast<Thread *>(parm);    Thread *myself = reinterpret_cast<Thread *>(parm);
    DQueue<MessageQueueService> *list = reinterpret_cast<DQueue<MessageQueueService> *>(myself->get_parm());      List<MessageQueueService, Mutex> *list =
    while (_stop_polling.value()  == 0)          reinterpret_cast<List<MessageQueueService, Mutex>*>(myself->get_parm());
   
       while (_stop_polling.get()  == 0)
    {    {
       _polling_sem.wait();       _polling_sem.wait();
  
       if (_stop_polling.value() != 0)          if (_stop_polling.get() != 0)
       {       {
          break;          break;
       }       }
  
           // The polling_routine thread must hold the lock on the
           // _polling_list while processing incoming messages.
           // This lock is used to give this thread ownership of
           // services on the _polling_routine list.
   
           // This is necessary to avoid confict with other threads
           // processing the _polling_list
           // (e.g., MessageQueueServer::~MessageQueueService).
   
       list->lock();       list->lock();
       int list_index = 0;          MessageQueueService *service = list->front();
       MessageQueueService *service = list->next(0);          ThreadStatus rtn = PEGASUS_THREAD_OK;
       while(service != NULL)       while(service != NULL)
         {         {
           ThreadStatus rtn;              if ((service->_incoming.count() > 0) &&
           rtn = PEGASUS_THREAD_OK;                  (service->_die.get() == 0) &&
           if (service->_incoming.count() > 0                  (service->_threads.get() < max_threads_per_svc_queue))
               && service->_die.value() == 0              {
               && service->_threads <= max_threads_per_svc_queue)                  // The _threads count is used to track the
             rtn = _thread_pool->allocate_and_awaken(service, _req_proc,                  // number of active threads that have been allocated
                                                         &_polling_sem);                  // to process messages for this service.
   
                   // The _threads count MUST be incremented while
                   // the polling_routine owns the _polling_thread
                   // lock and has ownership of the service object.
  
                   service->_threads++;
                   try
                   {
                       rtn = _thread_pool->allocate_and_awaken(
                           service, _req_proc, &_polling_sem);
                   }
                   catch (...)
                   {
                       service->_threads--;
   
                       // allocate_and_awaken should never generate an exception.
                       PEGASUS_ASSERT(0);
                   }
           // if no more threads available, break from processing loop           // if no more threads available, break from processing loop
           if (rtn != PEGASUS_THREAD_OK )           if (rtn != PEGASUS_THREAD_OK )
             {             {
                 Logger::put(Logger::STANDARD_LOG, System::CIMSERVER, Logger::TRACE,                      service->_threads--;
                         "Not enough threads to process this request. Skipping.");                      Logger::put(
                           Logger::STANDARD_LOG, System::CIMSERVER, Logger::TRACE,
                           "Not enough threads to process this request. "
                               "Skipping.");
  
                 Tracer::trace(TRC_MESSAGEQUEUESERVICE, Tracer::LEVEL2,                 Tracer::trace(TRC_MESSAGEQUEUESERVICE, Tracer::LEVEL2,
                         "Could not allocate thread for %s. " \                          "Could not allocate thread for %s.  Queue has %d "
                         "Queue has %d messages waiting and %s threads servicing." \                              "messages waiting and %d threads servicing."
                         "Skipping the service for right now. ",                         "Skipping the service for right now. ",
                         service->getQueueName(),                         service->getQueueName(),
                         service->_incoming.count(),                         service->_incoming.count(),
                         service->_threads.value());                          service->_threads.get());
  
               pegasus_yield();                      Threads::yield();
               service = NULL;               service = NULL;
             }             }
           else  
             {  
               service = list->next(service);  
             }  
         }         }
       list->unlock();              if (service != NULL)
   
       if (_check_idle_flag.value() != 0)  
       {       {
          _check_idle_flag = 0;                  service = list->next_of(service);
          // try to do idle thread clean up processing when system is not busy  
          // if system is busy there may not be a thread available to allocate  
          // so nothing will be done and that is OK.  
   
          if ( _thread_pool->allocate_and_awaken(service, kill_idle_threads, &_polling_sem) != PEGASUS_THREAD_OK)  
          {  
                 Logger::put(Logger::STANDARD_LOG, System::CIMSERVER, Logger::TRACE,  
                         "Not enough threads to kill idle threads. What an irony.");  
   
                 Tracer::trace(TRC_MESSAGEQUEUESERVICE, Tracer::LEVEL2,  
                         "Could not allocate thread to kill idle threads." \  
                         "Skipping. ");  
          }          }
   
   
       }       }
           list->unlock();
    }    }
    myself->exit_self( (PEGASUS_THREAD_RETURN) 1 );      myself->exit_self( (ThreadReturnType) 1 );
    return(0);      return 0;
 } }
  
  
 Semaphore MessageQueueService::_polling_sem(0); Semaphore MessageQueueService::_polling_sem(0);
 AtomicInt MessageQueueService::_stop_polling(0); AtomicInt MessageQueueService::_stop_polling(0);
 AtomicInt MessageQueueService::_check_idle_flag(0);  
  
  
 MessageQueueService::MessageQueueService( MessageQueueService::MessageQueueService(
Line 190 
Line 171 
      _mask(mask),      _mask(mask),
      _die(0),      _die(0),
         _threads(0),         _threads(0),
      _incoming(true, 0),        _incoming(),
      _incoming_queue_shutdown(0)      _incoming_queue_shutdown(0)
 { {
   
    _capabilities = (capabilities | module_capabilities::async);    _capabilities = (capabilities | module_capabilities::async);
  
    _default_op_timeout.tv_sec = 30;    _default_op_timeout.tv_sec = 30;
Line 201 
Line 181 
  
    max_threads_per_svc_queue = MAX_THREADS_PER_SVC_QUEUE;    max_threads_per_svc_queue = MAX_THREADS_PER_SVC_QUEUE;
  
    // if requested threads gt MAX_THREADS_PER_SVC_QUEUE_LIMIT      // if requested thread max is out of range, then set to
    // then set to MAX_THREADS_PER_SVC_QUEUE_LIMIT      // MAX_THREADS_PER_SVC_QUEUE_LIMIT
  
    if (max_threads_per_svc_queue > MAX_THREADS_PER_SVC_QUEUE_LIMIT)      if ((max_threads_per_svc_queue < 1) ||
           (max_threads_per_svc_queue > MAX_THREADS_PER_SVC_QUEUE_LIMIT))
      {      {
        max_threads_per_svc_queue = MAX_THREADS_PER_SVC_QUEUE_LIMIT;        max_threads_per_svc_queue = MAX_THREADS_PER_SVC_QUEUE_LIMIT;
      }      }
  
    // if requested threads eq 0 (unlimited)      Tracer::trace(TRC_MESSAGEQUEUESERVICE, Tracer::LEVEL2,
    // then set to MAX_THREADS_PER_SVC_QUEUE_LIMIT         "max_threads_per_svc_queue set to %u.", max_threads_per_svc_queue);
   
    if (max_threads_per_svc_queue == 0)  
      {  
        max_threads_per_svc_queue = MAX_THREADS_PER_SVC_QUEUE_DEFAULT;  
      }  
   
    // cout << "MAX_THREADS_PER_SVC_QUEUE = " << MAX_THREADS_PER_SVC_QUEUE << endl;  
    // cout << "max_threads_per_svc_queue set to = " << max_threads_per_svc_queue << endl;  
   
  
    AutoMutex autoMut(_meta_dispatcher_mutex);    AutoMutex autoMut(_meta_dispatcher_mutex);
  
    if (_meta_dispatcher == 0)    if (_meta_dispatcher == 0)
    {    {
       _stop_polling = 0;       _stop_polling = 0;
       PEGASUS_ASSERT(_service_count.value() == 0);          PEGASUS_ASSERT(_service_count.get() == 0);
       _meta_dispatcher = new cimom();       _meta_dispatcher = new cimom();
       if (_meta_dispatcher == NULL)  
       {  
          throw NullPointer();  
       }  
       //  _thread_pool = new ThreadPool(initial_cnt, "MessageQueueService",       //  _thread_pool = new ThreadPool(initial_cnt, "MessageQueueService",
       //   minimum_cnt, maximum_cnt, deallocateWait);       //   minimum_cnt, maximum_cnt, deallocateWait);
       //       //
Line 242 
Line 211 
  
    if (false == register_service(name, _capabilities, _mask))    if (false == register_service(name, _capabilities, _mask))
    {    {
       //l10n          MessageLoaderParms parms(
       //throw BindFailedException("MessageQueueService Base Unable to register with  Meta Dispatcher");              "Common.MessageQueueService.UNABLE_TO_REGISTER",
       MessageLoaderParms parms("Common.MessageQueueService.UNABLE_TO_REGISTER",              "CIM base message queue service is unable to register with the "
          "MessageQueueService Base Unable to register with  Meta Dispatcher");                  "CIMOM dispatcher.");
   
       throw BindFailedException(parms);       throw BindFailedException(parms);
    }    }
  
    _polling_list.insert_last(this);      _get_polling_list()->insert_back(this);
   
 } }
  
  
Line 259 
Line 226 
 { {
    _die = 1;    _die = 1;
  
    if (_incoming_queue_shutdown.value() == 0)      // The polling_routine locks the _polling_list while
       // processing the incoming messages for services on the
       // list.  Deleting the service from the _polling_list
       // prior to processing, avoids synchronization issues
       // with the _polling_routine.
   
       // ATTN: added to prevent assertion in List in which the list does not
       // contain this element.
   
       if (_get_polling_list()->contains(this))
           _get_polling_list()->remove(this);
   
       // ATTN: The code for closing the _incoming queue
       // is not working correctly. In OpenPegasus 2.5,
       // execution of the following code is very timing
       // dependent. This needs to be fix.
       // See Bug 4079 for details.
       if (_incoming_queue_shutdown.get() == 0)
    {    {
       _shutdown_incoming_queue();       _shutdown_incoming_queue();
   
    }    }
  
  while (_threads.value() > 0)      // Wait until all threads processing the messages
       // for this service have completed.
   
       while (_threads.get() > 0)
      {      {
           pegasus_yield();          Threads::yield();
      }      }
    _polling_list.remove(this);  
    {    {
      AutoMutex autoMut(_meta_dispatcher_mutex);      AutoMutex autoMut(_meta_dispatcher_mutex);
      _service_count--;      _service_count--;
      if (_service_count.value() == 0)          if (_service_count.get() == 0)
      {      {
  
       _stop_polling++;       _stop_polling++;
       _polling_sem.signal();       _polling_sem.signal();
       if (_polling_thread) {              if (_polling_thread)
               {
           _polling_thread->join();           _polling_thread->join();
           delete _polling_thread;           delete _polling_thread;
           _polling_thread = 0;           _polling_thread = 0;
Line 294 
Line 281 
    // Clean up in case there are extra stuff on the queue.    // Clean up in case there are extra stuff on the queue.
   while (_incoming.count())   while (_incoming.count())
   {   {
     try {          try
       delete _incoming.remove_first();          {
     } catch (const ListClosed &e)              delete _incoming.dequeue();
           }
           catch (const ListClosed&)
     {     {
       // If the list is closed, there is nothing we can do.       // If the list is closed, there is nothing we can do.
       break;       break;
Line 306 
Line 295 
  
 void MessageQueueService::_shutdown_incoming_queue() void MessageQueueService::_shutdown_incoming_queue()
 { {
    if (_incoming_queue_shutdown.value() > 0)      if (_incoming_queue_shutdown.get() > 0)
       return;       return;
  
    AsyncIoctl *msg = new AsyncIoctl(    AsyncIoctl *msg = new AsyncIoctl(
       get_next_xid(),  
       0,       0,
       _queueId,       _queueId,
       _queueId,       _queueId,
Line 326 
Line 314 
    msg->op->_state &= ~ASYNC_OPSTATE_COMPLETE;    msg->op->_state &= ~ASYNC_OPSTATE_COMPLETE;
  
    msg->op->_op_dest = this;    msg->op->_op_dest = this;
    msg->op->_request.insert_first(msg);      msg->op->_request.reset(msg);
    try {      try
      _incoming.insert_last_wait(msg->op);      {
           _incoming.enqueue_wait(msg->op);
      _polling_sem.signal();      _polling_sem.signal();
    } catch (const ListClosed &)      }
       catch (const ListClosed&)
    {    {
         // This means the queue has already been shut-down (happens  when there         // This means the queue has already been shut-down (happens  when there
     // are two AsyncIoctrl::IO_CLOSE messages generated and one got first     // are two AsyncIoctrl::IO_CLOSE messages generated and one got first
Line 355 
Line 345 
 } }
  
  
 PEGASUS_THREAD_RETURN PEGASUS_THREAD_CDECL MessageQueueService::_req_proc(  ThreadReturnType PEGASUS_THREAD_CDECL MessageQueueService::_req_proc(
     void * parm)     void * parm)
 { {
     MessageQueueService* service =     MessageQueueService* service =
Line 363 
Line 353 
     PEGASUS_ASSERT(service != 0);     PEGASUS_ASSERT(service != 0);
     try     try
     {     {
           if (service->_die.get() != 0)
         if (service->_die.value() != 0)  
         {         {
             return (0);              service->_threads--;
               return 0;
         }         }
             service->_threads++;  
         // pull messages off the incoming queue and dispatch them. then         // pull messages off the incoming queue and dispatch them. then
         // check pending messages that are non-blocking         // check pending messages that are non-blocking
         AsyncOpNode *operation = 0;         AsyncOpNode *operation = 0;
Line 378 
Line 367 
         {         {
             try             try
             {             {
                 operation = service->_incoming.remove_first();                  operation = service->_incoming.dequeue();
             }             }
             catch (ListClosed &)             catch (ListClosed &)
             {             {
Line 407 
Line 396 
             "Caught unrecognized exception.  Exiting _req_proc.");             "Caught unrecognized exception.  Exiting _req_proc.");
     }     }
     service->_threads--;     service->_threads--;
     return(0);      return 0;
 } }
  
  
Line 426 
Line 415 
 { {
    if (op->_flags & ASYNC_OPFLAGS_SAFE_CALLBACK)    if (op->_flags & ASYNC_OPFLAGS_SAFE_CALLBACK)
    {    {
           Message *msg = op->removeRequest();
       Message *msg = op->get_request();          if (msg && (msg->getMask() & MessageMask::ha_async))
       if (msg && (msg->getMask() & message_mask::ha_async))  
       {       {
          if (msg->getType() == async_messages::ASYNC_LEGACY_OP_START)          if (msg->getType() == async_messages::ASYNC_LEGACY_OP_START)
          {          {
Line 454 
Line 442 
          delete msg;          delete msg;
       }       }
  
       msg = op->get_response();          msg = op->removeResponse();
       if (msg && (msg->getMask() & message_mask::ha_async))          if (msg && (msg->getMask() & MessageMask::ha_async))
       {       {
          if (msg->getType() == async_messages::ASYNC_LEGACY_OP_RESULT)          if (msg->getType() == async_messages::ASYNC_LEGACY_OP_RESULT)
          {          {
Line 484 
Line 472 
       // note that _callback_node may be different from op       // note that _callback_node may be different from op
       // op->_callback_response_q is a "this" pointer we can use for       // op->_callback_response_q is a "this" pointer we can use for
       // static callback methods       // static callback methods
       op->_async_callback(op->_callback_node, op->_callback_response_q, op->_callback_ptr);          op->_async_callback(
               op->_callback_node, op->_callback_response_q, op->_callback_ptr);
    }    }
 } }
  
Line 498 
Line 487 
 // << Tue Feb 19 14:10:38 2002 mdd >> // << Tue Feb 19 14:10:38 2002 mdd >>
       operation->lock();       operation->lock();
  
       Message *rq = operation->_request.next(0);          Message *rq = operation->_request.get();
  
 // optimization <<< Thu Mar  7 21:04:05 2002 mdd >>> // optimization <<< Thu Mar  7 21:04:05 2002 mdd >>>
 // move this to the bottom of the loop when the majority of // move this to the bottom of the loop when the majority of
 // messages become async messages. // messages become async messages.
  
       // divert legacy messages to handleEnqueue       // divert legacy messages to handleEnqueue
       if ((rq != 0) && (!(rq->getMask() & message_mask::ha_async)))          if ((rq != 0) && (!(rq->getMask() & MessageMask::ha_async)))
       {       {
          rq = operation->_request.remove_first() ;              operation->_request.release();
          operation->unlock();          operation->unlock();
          // delete the op node          // delete the op node
          operation->release();          operation->release();
Line 568 
Line 557 
    Message* request,    Message* request,
    Message* response)    Message* response)
 { {
   
   STAT_COPYDISPATCHER  
   
    PEG_METHOD_ENTER(TRC_MESSAGEQUEUESERVICE,    PEG_METHOD_ENTER(TRC_MESSAGEQUEUESERVICE,
                     "MessageQueueService::_enqueueResponse");                     "MessageQueueService::_enqueueResponse");
  
    if (request->getMask() & message_mask::ha_async)      if (request->getMask() & MessageMask::ha_async)
    {    {
       if (response->getMask() & message_mask::ha_async)          if (response->getMask() & MessageMask::ha_async)
       {       {
          _completeAsyncResponse(static_cast<AsyncRequest *>(request),              _completeAsyncResponse(
                   static_cast<AsyncRequest *>(request),
                                 static_cast<AsyncReply *>(response),                                 static_cast<AsyncReply *>(response),
                                 ASYNC_OPSTATE_COMPLETE, 0);                                 ASYNC_OPSTATE_COMPLETE, 0);
          PEG_METHOD_EXIT();          PEG_METHOD_EXIT();
Line 589 
Line 576 
    if (request->_async != 0)    if (request->_async != 0)
    {    {
       Uint32 mask = request->_async->getMask();       Uint32 mask = request->_async->getMask();
       PEGASUS_ASSERT(mask & (message_mask::ha_async | message_mask::ha_request));          PEGASUS_ASSERT(mask &
               (MessageMask::ha_async | MessageMask::ha_request));
  
       AsyncRequest *async = static_cast<AsyncRequest *>(request->_async);       AsyncRequest *async = static_cast<AsyncRequest *>(request->_async);
       AsyncOpNode *op = async->op;       AsyncOpNode *op = async->op;
Line 601 
Line 589 
  
       AsyncLegacyOperationResult *async_result =       AsyncLegacyOperationResult *async_result =
          new AsyncLegacyOperationResult(          new AsyncLegacyOperationResult(
             async->getKey(),  
             async->getRouting(),  
             op,             op,
             response);             response);
       _completeAsyncResponse(       _completeAsyncResponse(
Line 617 
Line 603 
    // ensure that the destination queue is in response->dest    // ensure that the destination queue is in response->dest
    PEG_METHOD_EXIT();    PEG_METHOD_EXIT();
    return SendForget(response);    return SendForget(response);
   
 } }
  
 void MessageQueueService::_make_response(Message *req, Uint32 code) void MessageQueueService::_make_response(Message *req, Uint32 code)
Line 653 
Line 638 
  
 Boolean MessageQueueService::accept_async(AsyncOpNode *op) Boolean MessageQueueService::accept_async(AsyncOpNode *op)
 { {
    if (_incoming_queue_shutdown.value() > 0)      if (_incoming_queue_shutdown.get() > 0)
       return false;       return false;
    if (_polling_thread == NULL)    if (_polling_thread == NULL)
    {    {
       _polling_thread = new Thread(       _polling_thread = new Thread(
           polling_routine,           polling_routine,
           reinterpret_cast<void *>(&_polling_list),              reinterpret_cast<void *>(_get_polling_list()),
           false);           false);
       ThreadStatus tr = PEGASUS_THREAD_OK;       ThreadStatus tr = PEGASUS_THREAD_OK;
       while ( (tr =_polling_thread->run()) != PEGASUS_THREAD_OK)       while ( (tr =_polling_thread->run()) != PEGASUS_THREAD_OK)
       {       {
         if (tr == PEGASUS_THREAD_INSUFFICIENT_RESOURCES)         if (tr == PEGASUS_THREAD_INSUFFICIENT_RESOURCES)
            pegasus_yield();                  Threads::yield();
         else         else
            throw Exception(MessageLoaderParms("Common.MessageQueueService.NOT_ENOUGH_THREAD",                  throw Exception(MessageLoaderParms(
                       "Common.MessageQueueService.NOT_ENOUGH_THREAD",
                         "Could not allocate thread for the polling thread."));                         "Could not allocate thread for the polling thread."));
       }       }
    }    }
 // ATTN optimization remove the message checking altogether in the base // ATTN optimization remove the message checking altogether in the base
 // << Mon Feb 18 14:02:20 2002 mdd >> // << Mon Feb 18 14:02:20 2002 mdd >>
    op->lock();    op->lock();
    Message *rq = op->_request.next(0);      Message *rq = op->_request.get();
    Message *rp = op->_response.next(0);      Message *rp = op->_response.get();
    op->unlock();    op->unlock();
  
    if ((rq != 0 && (true == messageOK(rq))) ||    if ((rq != 0 && (true == messageOK(rq))) ||
        (rp != 0 && (true == messageOK(rp))) && _die.value() == 0)          (rp != 0 && (true == messageOK(rp))) && _die.get() == 0)
    {    {
       _incoming.insert_last_wait(op);          _incoming.enqueue_wait(op);
       _polling_sem.signal();       _polling_sem.signal();
       return true;       return true;
    }    }
Line 690 
Line 676 
  
 Boolean MessageQueueService::messageOK(const Message *msg) Boolean MessageQueueService::messageOK(const Message *msg)
 { {
    if (_incoming_queue_shutdown.value() > 0)      if (_incoming_queue_shutdown.get() > 0)
       return false;       return false;
    return true;    return true;
 } }
Line 701 
Line 687 
  
    AsyncReply *reply = new AsyncReply(    AsyncReply *reply = new AsyncReply(
       async_messages::HEARTBEAT,       async_messages::HEARTBEAT,
       req->getKey(),  
       req->getRouting(),  
       0,       0,
       req->op,       req->op,
       async_results::OK,       async_results::OK,
Line 722 
Line 706 
    {    {
       case AsyncIoctl::IO_CLOSE:       case AsyncIoctl::IO_CLOSE:
       {       {
          MessageQueueService *service = static_cast<MessageQueueService *>(req->op->_service_ptr);              MessageQueueService *service =
                   static_cast<MessageQueueService *>(req->op->_service_ptr);
  
 #ifdef MESSAGEQUEUESERVICE_DEBUG #ifdef MESSAGEQUEUESERVICE_DEBUG
          PEGASUS_STD(cout) << service->getQueueName() << " Received AsyncIoctl::IO_CLOSE " << PEGASUS_STD(endl);              PEGASUS_STD(cout) << service->getQueueName() <<
                   " Received AsyncIoctl::IO_CLOSE " << PEGASUS_STD(endl);
 #endif #endif
  
          // respond to this message. this is fire and forget, so we don't need to delete anything.              // respond to this message. this is fire and forget, so we
               // don't need to delete anything.
          // this takes care of two problems that were being found          // this takes care of two problems that were being found
          // << Thu Oct  9 10:52:48 2003 mdd >>          // << Thu Oct  9 10:52:48 2003 mdd >>
           _make_response(req, async_results::OK);           _make_response(req, async_results::OK);
          // ensure we do not accept any further messages          // ensure we do not accept any further messages
  
          // ensure we don't recurse on IO_CLOSE          // ensure we don't recurse on IO_CLOSE
          if (_incoming_queue_shutdown.value() > 0)              if (_incoming_queue_shutdown.get() > 0)
             break;             break;
  
          // set the closing flag          // set the closing flag
Line 746 
Line 733 
             AsyncOpNode *operation;             AsyncOpNode *operation;
             try             try
             {             {
                operation = service->_incoming.remove_first();                      operation = service->_incoming.dequeue();
             }             }
             catch(IPCException &)             catch(IPCException &)
             {             {
Line 761 
Line 748 
                break;                break;
          } // message processing loop          } // message processing loop
  
          // shutdown the AsyncDQueue              // shutdown the AsyncQueue
          service->_incoming.shutdown_queue();              service->_incoming.close();
          return;          return;
       }       }
  
Line 773 
Line 760 
  
 void MessageQueueService::handle_CimServiceStart(CimServiceStart *req) void MessageQueueService::handle_CimServiceStart(CimServiceStart *req)
 { {
   
 #ifdef MESSAGEQUEUESERVICE_DEBUG #ifdef MESSAGEQUEUESERVICE_DEBUG
    PEGASUS_STD(cout) << getQueueName() << "received START" << PEGASUS_STD(endl);      PEGASUS_STD(cout) << getQueueName() << "received START" <<
           PEGASUS_STD(endl);
 #endif #endif
  
    // clear the stoped bit and update    // clear the stoped bit and update
Line 783 
Line 770 
    _make_response(req, async_results::OK);    _make_response(req, async_results::OK);
    // now tell the meta dispatcher we are stopped    // now tell the meta dispatcher we are stopped
    update_service(_capabilities, _mask);    update_service(_capabilities, _mask);
   
 } }
   
 void MessageQueueService::handle_CimServiceStop(CimServiceStop *req) void MessageQueueService::handle_CimServiceStop(CimServiceStop *req)
 { {
 #ifdef MESSAGEQUEUESERVICE_DEBUG #ifdef MESSAGEQUEUESERVICE_DEBUG
Line 822 
Line 809 
  
 void MessageQueueService::handle_AsyncOperationResult(AsyncOperationResult *req) void MessageQueueService::handle_AsyncOperationResult(AsyncOperationResult *req)
 { {
    ;  
 } }
  
  
 void MessageQueueService::handle_AsyncLegacyOperationStart(AsyncLegacyOperationStart *req)  void MessageQueueService::handle_AsyncLegacyOperationStart(
       AsyncLegacyOperationStart* req)
 { {
    // remove the legacy message from the request and enqueue it to its destination      // remove the legacy message from the request and enqueue it to its
       // destination
    Uint32 result = async_results::CIM_NAK;    Uint32 result = async_results::CIM_NAK;
  
    Message *legacy = req->_act;    Message *legacy = req->_act;
Line 839 
Line 827 
       {       {
          if (queue->isAsync() == true)          if (queue->isAsync() == true)
          {          {
             (static_cast<MessageQueueService *>(queue))->handleEnqueue(legacy);                  (static_cast<MessageQueueService *>(queue))->handleEnqueue(
                       legacy);
          }          }
          else          else
          {          {
Line 853 
Line 842 
    _make_response(req, result);    _make_response(req, result);
 } }
  
 void MessageQueueService::handle_AsyncLegacyOperationResult(AsyncLegacyOperationResult *rep)  void MessageQueueService::handle_AsyncLegacyOperationResult(
       AsyncLegacyOperationResult* rep)
 { {
    ;  
 } }
  
 AsyncOpNode *MessageQueueService::get_op() AsyncOpNode *MessageQueueService::get_op()
Line 870 
Line 859 
  
 void MessageQueueService::return_op(AsyncOpNode *op) void MessageQueueService::return_op(AsyncOpNode *op)
 { {
    PEGASUS_ASSERT(op->read_state() & ASYNC_OPSTATE_RELEASED);      PEGASUS_ASSERT(op->_state & ASYNC_OPSTATE_RELEASED);
    delete op;    delete op;
 } }
  
  
 Boolean MessageQueueService::ForwardOp(  
     AsyncOpNode *op,  
     Uint32 destination)  
 {  
    PEGASUS_ASSERT(op != 0);  
    op->lock();  
    op->_op_dest = MessageQueue::lookup(destination);  
    op->_flags |= (ASYNC_OPFLAGS_FIRE_AND_FORGET | ASYNC_OPFLAGS_FORWARD);  
    op->_flags &= ~(ASYNC_OPFLAGS_CALLBACK);  
    op->unlock();  
    if (op->_op_dest == 0)  
       return false;  
   
    return  _meta_dispatcher->route_async(op);  
 }  
   
   
 Boolean MessageQueueService::SendAsync( Boolean MessageQueueService::SendAsync(
     AsyncOpNode *op,     AsyncOpNode *op,
     Uint32 destination,     Uint32 destination,
Line 904 
Line 876 
    // get the queue handle for the destination    // get the queue handle for the destination
  
    op->lock();    op->lock();
    op->_op_dest = MessageQueue::lookup(destination); // destination of this message      // destination of this message
       op->_op_dest = MessageQueue::lookup(destination);
    op->_flags |= ASYNC_OPFLAGS_CALLBACK;    op->_flags |= ASYNC_OPFLAGS_CALLBACK;
    op->_flags &= ~(ASYNC_OPFLAGS_FIRE_AND_FORGET);    op->_flags &= ~(ASYNC_OPFLAGS_FIRE_AND_FORGET);
    // initialize the callback data    // initialize the callback data
    op->_async_callback = callback;   // callback function to be executed by recpt. of response      // callback function to be executed by recpt. of response
    op->_callback_node = op;          // the op node      op->_async_callback = callback;
    op->_callback_response_q = callback_response_q;  // the queue that will receive the response      // the op node
    op->_callback_ptr = callback_ptr;   // user data for callback      op->_callback_node = op;
    op->_callback_request_q = this;     // I am the originator of this request      // the queue that will receive the response
       op->_callback_response_q = callback_response_q;
       // user data for callback
       op->_callback_ptr = callback_ptr;
       // I am the originator of this request
       op->_callback_request_q = this;
  
    op->unlock();    op->unlock();
    if (op->_op_dest == 0)    if (op->_op_dest == 0)
Line 950 
Line 928 
    op->_callback_parameter = parameter;    op->_callback_parameter = parameter;
    op->_callback_response_q = this;    op->_callback_response_q = this;
  
    if (!(msg->getMask() & message_mask::ha_async))      if (!(msg->getMask() & MessageMask::ha_async))
    {    {
       AsyncLegacyOperationStart *wrapper = new AsyncLegacyOperationStart(       AsyncLegacyOperationStart *wrapper = new AsyncLegacyOperationStart(
          get_next_xid(),  
          op,          op,
          destination,          destination,
          msg,          msg,
Line 961 
Line 938 
    }    }
    else    else
    {    {
       op->_request.insert_first(msg);          op->_request.reset(msg);
       (static_cast<AsyncMessage *>(msg))->op = op;       (static_cast<AsyncMessage *>(msg))->op = op;
    }    }
    return _meta_dispatcher->route_async(op);    return _meta_dispatcher->route_async(op);
Line 973 
Line 950 
    AsyncOpNode *op = 0;    AsyncOpNode *op = 0;
    Uint32 mask = msg->getMask();    Uint32 mask = msg->getMask();
  
    if (mask & message_mask::ha_async)      if (mask & MessageMask::ha_async)
    {    {
       op = (static_cast<AsyncMessage *>(msg))->op ;       op = (static_cast<AsyncMessage *>(msg))->op ;
    }    }
Line 981 
Line 958 
    if (op == 0)    if (op == 0)
    {    {
       op = get_op();       op = get_op();
       op->_request.insert_first(msg);          op->_request.reset(msg);
       if (mask & message_mask::ha_async)          if (mask & MessageMask::ha_async)
       {       {
          (static_cast<AsyncMessage *>(msg))->op = op;          (static_cast<AsyncMessage *>(msg))->op = op;
       }       }
Line 1014 
Line 991 
    if (request->op == 0)    if (request->op == 0)
    {    {
       request->op = get_op();       request->op = get_op();
       request->op->_request.insert_first(request);          request->op->_request.reset(request);
       destroy_op = true;       destroy_op = true;
    }    }
  
Line 1029 
Line 1006 
  
    request->op->_client_sem.wait();    request->op->_client_sem.wait();
  
    request->op->lock();      AsyncReply* rpl = static_cast<AsyncReply *>(request->op->removeResponse());
    AsyncReply * rpl = static_cast<AsyncReply *>(request->op->_response.remove_first());  
    rpl->op = 0;    rpl->op = 0;
    request->op->unlock();  
  
    if (destroy_op == true)    if (destroy_op == true)
    {    {
       request->op->lock();       request->op->lock();
       request->op->_request.remove(request);          request->op->_request.release();
       request->op->_state |= ASYNC_OPSTATE_RELEASED;       request->op->_state |= ASYNC_OPSTATE_RELEASED;
       request->op->unlock();       request->op->unlock();
       return_op(request->op);       return_op(request->op);
Line 1053 
Line 1028 
     Uint32 mask)     Uint32 mask)
 { {
    RegisterCimService *msg = new RegisterCimService(    RegisterCimService *msg = new RegisterCimService(
       get_next_xid(),  
       0,       0,
       true,       true,
       name,       name,
Line 1067 
Line 1041 
  
    if (reply != 0)    if (reply != 0)
    {    {
       if (reply->getMask() & message_mask::ha_async)          if (reply->getMask() & MessageMask::ha_async)
       {       {
          if (reply->getMask() & message_mask::ha_reply)              if (reply->getMask() & MessageMask::ha_reply)
          {          {
             if (reply->result == async_results::OK ||             if (reply->result == async_results::OK ||
                 reply->result == async_results::MODULE_ALREADY_REGISTERED)                 reply->result == async_results::MODULE_ALREADY_REGISTERED)
Line 1088 
Line 1062 
 Boolean MessageQueueService::update_service(Uint32 capabilities, Uint32 mask) Boolean MessageQueueService::update_service(Uint32 capabilities, Uint32 mask)
 { {
    UpdateCimService *msg = new UpdateCimService(    UpdateCimService *msg = new UpdateCimService(
       get_next_xid(),  
       0,       0,
       true,       true,
       _queueId,       _queueId,
Line 1099 
Line 1072 
    AsyncMessage *reply = SendWait(msg);    AsyncMessage *reply = SendWait(msg);
    if (reply)    if (reply)
    {    {
       if (reply->getMask() & message_mask::ha_async)          if (reply->getMask() & MessageMask::ha_async)
       {       {
          if (reply->getMask() & message_mask::ha_reply)              if (reply->getMask() & MessageMask::ha_reply)
          {          {
             if (static_cast<AsyncReply *>(reply)->result == async_results::OK)                  if (static_cast<AsyncReply *>(reply)->result ==
                           async_results::OK)
             {             {
                registered = true;                registered = true;
             }             }
Line 1118 
Line 1092 
  
 Boolean MessageQueueService::deregister_service() Boolean MessageQueueService::deregister_service()
 { {
   
    _meta_dispatcher->deregister_module(_queueId);    _meta_dispatcher->deregister_module(_queueId);
    return true;    return true;
 } }
Line 1138 
Line 1111 
    results->clear();    results->clear();
  
    FindServiceQueue *req = new FindServiceQueue(    FindServiceQueue *req = new FindServiceQueue(
       get_next_xid(),  
       0,       0,
       _queueId,       _queueId,
       true,       true,
Line 1151 
Line 1123 
    AsyncMessage *reply = SendWait(req);    AsyncMessage *reply = SendWait(req);
    if (reply)    if (reply)
    {    {
       if (reply->getMask() & message_mask::ha_async)          if (reply->getMask() & MessageMask::ha_async)
       {       {
          if (reply->getMask() & message_mask::ha_reply)              if (reply->getMask() & MessageMask::ha_reply)
          {          {
             if (reply->getType() == async_messages::FIND_SERVICE_Q_RESULT)             if (reply->getType() == async_messages::FIND_SERVICE_Q_RESULT)
             {             {
                if ((static_cast<FindServiceQueueResult *>(reply))->result == async_results::OK)                      if ((static_cast<FindServiceQueueResult*>(reply))->result ==
                   *results = (static_cast<FindServiceQueueResult *>(reply))->qids;                              async_results::OK)
                           *results =
                               (static_cast<FindServiceQueueResult*>(reply))->qids;
             }             }
          }          }
       }       }
       delete reply;       delete reply;
    }    }
    delete req;    delete req;
    return ;  
 } }
  
 void MessageQueueService::enumerate_service(Uint32 queue, message_module *result)  void MessageQueueService::enumerate_service(
       Uint32 queue,
       message_module* result)
 { {
    if (result == 0)    if (result == 0)
    {    {
Line 1176 
Line 1151 
    }    }
  
    EnumerateService *req = new EnumerateService(    EnumerateService *req = new EnumerateService(
       get_next_xid(),  
       0,       0,
       _queueId,       _queueId,
       true,       true,
Line 1188 
Line 1162 
    {    {
       Boolean found = false;       Boolean found = false;
  
       if (reply->getMask() & message_mask::ha_async)          if (reply->getMask() & MessageMask::ha_async)
       {       {
          if (reply->getMask() & message_mask::ha_reply)              if (reply->getMask() & MessageMask::ha_reply)
          {          {
             if (reply->getType() == async_messages::ENUMERATE_SERVICE_RESULT)                  if (reply->getType() ==
                           async_messages::ENUMERATE_SERVICE_RESULT)
             {             {
                if ((static_cast<EnumerateServiceResponse *>(reply))->result == async_results::OK)                      if ((static_cast<EnumerateServiceResponse*>(reply))->
                               result == async_results::OK)
                {                {
                   if (found == false)                   if (found == false)
                   {                   {
                      found = true;                      found = true;
  
                      result->put_name((static_cast<EnumerateServiceResponse *>(reply))->name);                              result->put_name((static_cast<
                      result->put_capabilities((static_cast<EnumerateServiceResponse *>(reply))->capabilities);                                  EnumerateServiceResponse*>(reply))->name);
                      result->put_mask((static_cast<EnumerateServiceResponse *>(reply))->mask);                              result->put_capabilities((static_cast<
                      result->put_queue((static_cast<EnumerateServiceResponse *>(reply))->qid);                                  EnumerateServiceResponse*>(reply))->
                                       capabilities);
                               result->put_mask((static_cast<
                                   EnumerateServiceResponse*>(reply))->mask);
                               result->put_queue((static_cast<
                                   EnumerateServiceResponse*>(reply))->qid);
                   }                   }
                }                }
             }             }
Line 1212 
Line 1193 
       delete reply;       delete reply;
    }    }
    delete req;    delete req;
   
    return;  
 } }
  
 Uint32 MessageQueueService::get_next_xid()  MessageQueueService::PollingList* MessageQueueService::_get_polling_list()
 { {
    static Mutex _monitor;      _polling_list_mutex.lock();
    Uint32 value;  
    AutoMutex autoMut(_monitor);      if (!_polling_list)
    _xid++;          _polling_list = new PollingList;
    value =  _xid.value();  
    return value;      _polling_list_mutex.unlock();
  
       return _polling_list;
 } }
  
 PEGASUS_NAMESPACE_END PEGASUS_NAMESPACE_END


Legend:
Removed from v.1.114  
changed lines
  Added in v.1.131

No CVS admin address has been configured
Powered by
ViewCVS 0.9.2