(file) Return to MessageQueueService.cpp CVS log (file) (dir) Up to [Pegasus] / pegasus / src / Pegasus / Common

Diff for /pegasus/src/Pegasus/Common/MessageQueueService.cpp between version 1.4 and 1.151

version 1.4, 2002/01/21 02:01:22 version 1.151, 2008/11/05 05:24:34
Line 1 
Line 1 
 //%////-*-c++-*-////////////////////////////////////////////////////////////////  //%2006////////////////////////////////////////////////////////////////////////
 // //
 // Copyright (c) 2000, 2001 The Open group, BMC Software, Tivoli Systems, IBM  // Copyright (c) 2000, 2001, 2002 BMC Software; Hewlett-Packard Development
   // Company, L.P.; IBM Corp.; The Open Group; Tivoli Systems.
   // Copyright (c) 2003 BMC Software; Hewlett-Packard Development Company, L.P.;
   // IBM Corp.; EMC Corporation, The Open Group.
   // Copyright (c) 2004 BMC Software; Hewlett-Packard Development Company, L.P.;
   // IBM Corp.; EMC Corporation; VERITAS Software Corporation; The Open Group.
   // Copyright (c) 2005 Hewlett-Packard Development Company, L.P.; IBM Corp.;
   // EMC Corporation; VERITAS Software Corporation; The Open Group.
   // Copyright (c) 2006 Hewlett-Packard Development Company, L.P.; IBM Corp.;
   // EMC Corporation; Symantec Corporation; The Open Group.
 // //
 // Permission is hereby granted, free of charge, to any person obtaining a copy // Permission is hereby granted, free of charge, to any person obtaining a copy
 // of this software and associated documentation files (the "Software"), to // of this software and associated documentation files (the "Software"), to
Line 20 
Line 29 
 // //
 //============================================================================== //==============================================================================
 // //
 // Author: Mike Day (mdday@us.ibm.com)  
 //  
 // Modified By:  
 //  
 //%///////////////////////////////////////////////////////////////////////////// //%/////////////////////////////////////////////////////////////////////////////
  
 #include "MessageQueueService.h" #include "MessageQueueService.h"
   #include <Pegasus/Common/Tracer.h>
   #include <Pegasus/Common/MessageLoader.h>
  
 PEGASUS_NAMESPACE_BEGIN PEGASUS_NAMESPACE_BEGIN
  
   cimom *MessageQueueService::_meta_dispatcher = 0;
   AtomicInt MessageQueueService::_service_count(0);
   Mutex MessageQueueService::_meta_dispatcher_mutex;
  
 PEGASUS_THREAD_RETURN PEGASUS_THREAD_CDECL MessageQueueService::_req_proc(void * parm)  static struct timeval deallocateWait = {300, 0};
 {  
    Thread *myself = reinterpret_cast<Thread *>(parm);  
    MessageQueueService *service = reinterpret_cast<MessageQueueService *>(myself->get_parm());  
  
    // pull messages off the incoming queue and dispatch them. then  ThreadPool *MessageQueueService::_thread_pool = 0;
    // check pending messages that are non-blocking  
  
    while ( service->_die.value() < 1 )  MessageQueueService::PollingList* MessageQueueService::_polling_list;
    {  Mutex MessageQueueService::_polling_list_mutex;
  
       AsyncOpNode *operation = service->_incoming.remove_first_wait();  Thread* MessageQueueService::_polling_thread = 0;
       service->_handle_incoming_operation(operation);  
    }  
  
    myself->exit_self( (PEGASUS_THREAD_RETURN) 1 );  ThreadPool *MessageQueueService::get_thread_pool()
    return(0);  {
      return _thread_pool;
 } }
  
 PEGASUS_THREAD_RETURN PEGASUS_THREAD_CDECL MessageQueueService::_rpl_proc(void * parm)  //
   // MAX_THREADS_PER_SVC_QUEUE
   //
   // JR Wunderlich Jun 6, 2005
   //
   
   #define MAX_THREADS_PER_SVC_QUEUE_LIMIT 5000
   #define MAX_THREADS_PER_SVC_QUEUE_DEFAULT 5
   
   #ifndef MAX_THREADS_PER_SVC_QUEUE
   # define MAX_THREADS_PER_SVC_QUEUE MAX_THREADS_PER_SVC_QUEUE_DEFAULT
   #endif
   
   Uint32 max_threads_per_svc_queue;
   
   ThreadReturnType PEGASUS_THREAD_CDECL MessageQueueService::polling_routine(
       void* parm)
 { {
    Thread *myself = reinterpret_cast<Thread *>(parm);    Thread *myself = reinterpret_cast<Thread *>(parm);
    MessageQueueService *service = reinterpret_cast<MessageQueueService *>(myself->get_parm());      List<MessageQueueService, Mutex> *list =
           reinterpret_cast<List<MessageQueueService, Mutex>*>(myself->get_parm());
    // pull messages off the _pending queue and complete them.  
    // this loop is not optimized because every time we complete an operation  
    // we start again to traverse the loop at the beginning.  
    Uint32 loop = 1;  
    Boolean unlinked = false;  
  
       while (_stop_polling.get()  == 0)
       {
           _polling_sem.wait();
  
    while ( service->_die.value() < 1 )          if (_stop_polling.get() != 0)
    {    {
       service->_pending.wait_for_node();              break;
       AsyncOpNode *operation = service->_pending.next(0);          }
 //      operation = service->_pending.remove_no_lock(operation);  
       while( operation != 0 )          // The polling_routine thread must hold the lock on the
           // _polling_list while processing incoming messages.
           // This lock is used to give this thread ownership of
           // services on the _polling_routine list.
   
           // This is necessary to avoid confict with other threads
           // processing the _polling_list
           // (e.g., MessageQueueServer::~MessageQueueService).
   
           list->lock();
           MessageQueueService *service = list->front();
           ThreadStatus rtn = PEGASUS_THREAD_OK;
           while (service != NULL)
       {       {
          if ( operation->_user_data != loop )              if ((service->_incoming.count() > 0) &&
                   (service->_die.get() == 0) &&
                   (service->_threads.get() < max_threads_per_svc_queue))
          {          {
             Uint32 state = operation->read_state();                  // The _threads count is used to track the
                   // number of active threads that have been allocated
                   // to process messages for this service.
   
                   // The _threads count MUST be incremented while
                   // the polling_routine owns the _polling_thread
                   // lock and has ownership of the service object.
  
             if ( state  & ASYNC_OPSTATE_COMPLETE )                  service->_threads++;
                   try
             {             {
                service->_pending.remove_no_lock(operation) ;                      rtn = _thread_pool->allocate_and_awaken(
                operation->_client_sem.signal();                          service, _req_proc, &_polling_sem);
                operation = service->_pending.next(0);  
                unlinked = true;  
                continue;  
             }             }
             operation->_user_data = loop;                  catch (...)
                   {
                       service->_threads--;
   
                       // allocate_and_awaken should never generate an exception.
                       PEGASUS_ASSERT(0);
                   }
                   // if no more threads available, break from processing loop
                   if (rtn != PEGASUS_THREAD_OK )
                   {
                       service->_threads--;
                       PEG_TRACE((TRC_MESSAGEQUEUESERVICE, Tracer::LEVEL1,
                           "Could not allocate thread for %s.  Queue has %d "
                               "messages waiting and %d threads servicing."
                               "Skipping the service for right now. ",
                           service->getQueueName(),
                           service->_incoming.count(),
                           service->_threads.get()));
   
                       Threads::yield();
                       service = NULL;
          }          }
          operation = service->_pending.next(operation);  
       }       }
       service->_pending.unlock();              if (service != NULL)
       loop++;  
       if ( unlinked == true)  
       {       {
          pegasus_yield();                  service = list->next_of(service);
          unlinked = false;  
       }       }
    }    }
           list->unlock();
    myself->exit_self( (PEGASUS_THREAD_RETURN) 1 );      }
    return(0);      return ThreadReturnType(0);
 } }
  
  
 void MessageQueueService::_handle_incoming_operation(AsyncOpNode *operation)  Semaphore MessageQueueService::_polling_sem(0);
   AtomicInt MessageQueueService::_stop_polling(0);
   
   
   MessageQueueService::MessageQueueService(
       const char* name,
       Uint32 queueID)
       : Base(name, true,  queueID),
         _die(0),
         _threads(0),
         _incoming(),
         _incoming_queue_shutdown(0)
 { {
    if ( operation != 0 )      _isRunning = true;
   
       max_threads_per_svc_queue = MAX_THREADS_PER_SVC_QUEUE;
   
       // if requested thread max is out of range, then set to
       // MAX_THREADS_PER_SVC_QUEUE_LIMIT
   
       if ((max_threads_per_svc_queue < 1) ||
           (max_threads_per_svc_queue > MAX_THREADS_PER_SVC_QUEUE_LIMIT))
    {    {
       Message *rq = operation->get_request();          max_threads_per_svc_queue = MAX_THREADS_PER_SVC_QUEUE_LIMIT;
       PEGASUS_ASSERT(rq != 0 );  
       PEGASUS_ASSERT(rq->getMask() & message_mask::ha_async );  
       PEGASUS_ASSERT(rq->getMask() & message_mask::ha_request);  
       _handle_async_request(static_cast<AsyncRequest *>(rq));  
    }    }
  
    return;      PEG_TRACE((TRC_MESSAGEQUEUESERVICE, Tracer::LEVEL3,
          "max_threads_per_svc_queue set to %u.", max_threads_per_svc_queue));
  
       AutoMutex autoMut(_meta_dispatcher_mutex);
   
       if (_meta_dispatcher == 0)
       {
           _stop_polling = 0;
           PEGASUS_ASSERT(_service_count.get() == 0);
           _meta_dispatcher = new cimom();
   
           //  _thread_pool = new ThreadPool(initial_cnt, "MessageQueueService",
           //   minimum_cnt, maximum_cnt, deallocateWait);
           //
           _thread_pool =
               new ThreadPool(0, "MessageQueueService", 0, 0, deallocateWait);
       }
       _service_count++;
   
       _get_polling_list()->insert_back(this);
 } }
  
  
 Boolean MessageQueueService::messageOK(const Message *msg)  MessageQueueService::~MessageQueueService()
 { {
    if ( msg != 0 )      _die = 1;
   
       // The polling_routine locks the _polling_list while
       // processing the incoming messages for services on the
       // list.  Deleting the service from the _polling_list
       // prior to processing, avoids synchronization issues
       // with the _polling_routine.
   
       // ATTN: added to prevent assertion in List in which the list does not
       // contain this element.
   
       if (_get_polling_list()->contains(this))
           _get_polling_list()->remove(this);
   
       // ATTN: The code for closing the _incoming queue
       // is not working correctly. In OpenPegasus 2.5,
       // execution of the following code is very timing
       // dependent. This needs to be fix.
       // See Bug 4079 for details.
       if (_incoming_queue_shutdown.get() == 0)
    {    {
       Uint32 mask = msg->getMask();          _shutdown_incoming_queue();
       if ( mask & message_mask::ha_async)  
          if ( mask & message_mask::ha_request)  
             return true;  
    }  
    return false;  
 } }
  
       // Wait until all threads processing the messages
       // for this service have completed.
  
 MessageQueueService::MessageQueueService(const char *name,      while (_threads.get() > 0)
                                          Uint32 queueID,      {
                                          Uint32 capabilities,          Threads::yield();
                                          Uint32 mask)  
    : Base(name, false,  queueID),  
      _capabilities(capabilities),  
      _mask(mask),  
      _die(0),  
      _pending(true, 1000),  
      _incoming(true, 1000),  
      _req_thread(_req_proc, this, false),  
      _rpl_thread(_rpl_proc, this, false)  
 {  
    _default_op_timeout.tv_sec = 30;  
    _default_op_timeout.tv_usec = 100;  
    _meta_dispatcher = static_cast<cimom *>(Base::lookup(CIMOM_Q_ID));  
    if(_meta_dispatcher == 0 )  
       throw NullPointer();  
    _req_thread.run();  
    _rpl_thread.run();  
   
 } }
  
       {
           AutoMutex autoMut(_meta_dispatcher_mutex);
           _service_count--;
           if (_service_count.get() == 0)
           {
  
 MessageQueueService::~MessageQueueService(void)              _stop_polling++;
               _polling_sem.signal();
               if (_polling_thread)
 { {
    _die = 1;                  _polling_thread->join();
    _pending.shutdown_queue();                  delete _polling_thread;
    _incoming.shutdown_queue();                  _polling_thread = 0;
    _req_thread.join();  
    _rpl_thread.join();  
 } }
               _meta_dispatcher->_shutdown_routed_queue();
               delete _meta_dispatcher;
               _meta_dispatcher = 0;
  
 AtomicInt MessageQueueService::_xid(1);              delete _thread_pool;
               _thread_pool = 0;
           }
       } // mutex unlocks here
  
       // Clean up any extra stuff on the queue.
       AsyncOpNode* op = 0;
       while ((op = _incoming.dequeue()))
       {
           delete op;
       }
   }
  
 Boolean MessageQueueService::accept_async(AsyncOpNode *op) throw(IPCException)  void MessageQueueService::_shutdown_incoming_queue()
 { {
    // incoming async request message      if (_incoming_queue_shutdown.get() > 0)
           return;
  
    PEGASUS_ASSERT(op != 0 );      AsyncIoctl *msg = new AsyncIoctl(
           0,
           _queueId,
           _queueId,
           true,
           AsyncIoctl::IO_CLOSE,
           0,
           0);
  
    Message *rq = op->get_request();      msg->op = get_op();
    if ( true == messageOK(rq) )      msg->op->_flags = ASYNC_OPFLAGS_FIRE_AND_FORGET;
    {  
       try      msg->op->_op_dest = this;
       msg->op->_request.reset(msg);
       if (_incoming.enqueue(msg->op))
       {       {
          _incoming.insert_first_wait(op);          _polling_sem.signal();
       }       }
       catch ( IPCException& e)      else
       {       {
          return false ;          // This means the queue has already been shut-down (happens  when there
       }          // are two AsyncIoctrl::IO_CLOSE messages generated and one got first
       return true;          // processed.
           delete msg;
    }    }
    return false;  
 } }
  
  
 void MessageQueueService::_completeAsyncResponse(AsyncRequest *request,  void MessageQueueService::enqueue(Message* msg)
                                                 AsyncReply *reply,  
                                                 Uint32 state,  
                                                 Uint32 flag)  
 { {
    PEGASUS_ASSERT(request != 0  && reply != 0 );      PEG_METHOD_ENTER(TRC_MESSAGEQUEUESERVICE, "MessageQueueService::enqueue()");
  
    AsyncOpNode *op = request->op;      Base::enqueue(msg);
    op->lock();  
 //   if (false == op->_response.exists(reply))  
 //      op->_response.insert_last(reply);  
    op->_response = reply;  
  
    op->_state |= state ;      PEG_METHOD_EXIT();
    op->_flags |= flag;  
    gettimeofday(&(op->_updated), NULL);  
    op->unlock();  
 } }
  
 void MessageQueueService::handleEnqueue(void)  
   ThreadReturnType PEGASUS_THREAD_CDECL MessageQueueService::_req_proc(
       void* parm)
 { {
    Message *msg = dequeue();      MessageQueueService* service =
    if( msg )          reinterpret_cast<MessageQueueService*>(parm);
       PEGASUS_ASSERT(service != 0);
       try
    {    {
       delete msg;          if (service->_die.get() != 0)
    }          {
               service->_threads--;
               return 0;
 } }
           // pull messages off the incoming queue and dispatch them. then
           // check pending messages that are non-blocking
           AsyncOpNode *operation = 0;
  
 void MessageQueueService::_handle_async_request(AsyncRequest *req)          // many operations may have been queued.
           do
 { {
    if ( req != 0 )              operation = service->_incoming.dequeue();
    {  
       req->op->processing();  
  
       Uint32 type = req->getType();              if (operation)
       if( type == async_messages::HEARTBEAT )  
          handle_heartbeat_request(req);  
       else if (type == async_messages::IOCTL)  
          handle_AsyncIoctl(static_cast<AsyncIoctl *>(req));  
       else if (type == async_messages::CIMSERVICE_START)  
          handle_CimServiceStart(static_cast<CimServiceStart *>(req));  
       else if (type == async_messages::CIMSERVICE_STOP)  
          handle_CimServiceStop(static_cast<CimServiceStop *>(req));  
       else if (type == async_messages::CIMSERVICE_PAUSE)  
          handle_CimServicePause(static_cast<CimServicePause *>(req));  
       else if (type == async_messages::CIMSERVICE_RESUME)  
          handle_CimServiceResume(static_cast<CimServiceResume *>(req));  
       else if ( type == async_messages::ASYNC_OP_START)  
          handle_AsyncOperationStart(static_cast<AsyncOperationStart *>(req));  
       else  
       {       {
          // we don't handle this request message                 operation->_service_ptr = service;
          _make_response(req, async_results::CIM_NAK );                 service->_handle_incoming_operation(operation);
       }       }
           } while (operation);
    }    }
       catch (const Exception& e)
       {
           PEG_TRACE((TRC_DISCARDED_DATA, Tracer::LEVEL1,
               "Caught exception: \"%s\".  Exiting _req_proc.",
               (const char*)e.getMessage().getCString()));
 } }
       catch (...)
 void MessageQueueService::_handle_async_reply(AsyncReply *rep)  
 { {
           PEG_TRACE_CSTRING(TRC_DISCARDED_DATA, Tracer::LEVEL1,
               "Caught unrecognized exception.  Exiting _req_proc.");
       }
       service->_threads--;
       return 0;
   }
  
    if (rep->op != 0 )  
       rep->op->processing();  
   
    Uint32 type = rep->getType();  
  
    if ( type == async_messages::ASYNC_OP_RESULT )  void MessageQueueService::_sendwait_callback(
       handle_AsyncOperationResult(static_cast<AsyncOperationResult *>(rep));      AsyncOpNode* op,
    else      MessageQueue* q,
       void *parm)
    {    {
       // we don't handle this reply      op->_client_sem.signal();
       ;  
    }    }
  
    if( rep->op != 0 )  
       rep->op->release();  
 }  
  
 void MessageQueueService::_make_response(AsyncRequest *req, Uint32 code)  // callback function is responsible for cleaning up all resources
   // including op, op->_callback_node, and op->_callback_ptr
   void MessageQueueService::_handle_async_callback(AsyncOpNode* op)
 { {
    AsyncReply *reply =      PEGASUS_ASSERT(op->_flags == ASYNC_OPFLAGS_CALLBACK);
       new AsyncReply(async_messages::REPLY,      // note that _callback_node may be different from op
                      req->getKey(),      // op->_callback_response_q is a "this" pointer we can use for
                      req->getRouting(),      // static callback methods
                      0,      op->_async_callback(
                      req->op,          op->_callback_node, op->_callback_response_q, op->_callback_ptr);
                      code,  
                      req->resp,  
                      false);  
    _completeAsyncResponse(req, reply, ASYNC_OPSTATE_COMPLETE, 0 );  
 } }
  
  
 void MessageQueueService::handle_heartbeat_request(AsyncRequest *req)  void MessageQueueService::_handle_incoming_operation(AsyncOpNode* operation)
 { {
    // default action is to echo a heartbeat response      if (operation != 0)
       {
           Message *rq = operation->_request.get();
  
    AsyncReply *reply =  // optimization <<< Thu Mar  7 21:04:05 2002 mdd >>>
       new AsyncReply(async_messages::HEARTBEAT,  // move this to the bottom of the loop when the majority of
                      req->getKey(),  // messages become async messages.
                      req->getRouting(),  
                      0,          // divert legacy messages to handleEnqueue
                      req->op,          if ((rq != 0) && (!(rq->getMask() & MessageMask::ha_async)))
                      async_results::OK,          {
                      req->resp,              operation->_request.release();
                      false);              // delete the op node
    _completeAsyncResponse(req, reply, ASYNC_OPSTATE_COMPLETE, 0 );              return_op(operation);
               handleEnqueue(rq);
               return;
 } }
  
           if ((operation->_flags & ASYNC_OPFLAGS_CALLBACK) &&
 void MessageQueueService::handle_heartbeat_reply(AsyncReply *rep)              (operation->_state & ASYNC_OPSTATE_COMPLETE))
 { {
    ;              _handle_async_callback(operation);
 } }
           else
 void MessageQueueService::handle_AsyncIoctl(AsyncIoctl *req)  
 { {
    _make_response(req, async_results::OK);              PEGASUS_ASSERT(rq != 0);
               _handle_async_request(static_cast<AsyncRequest *>(rq));
 } }
 void MessageQueueService::handle_CimServiceStart(CimServiceStart *req)  
 {  
    _make_response(req, async_results::CIM_NAK);  
 } }
 void MessageQueueService::handle_CimServiceStop(CimServiceStop *req)      return;
   }
   
   void MessageQueueService::_handle_async_request(AsyncRequest *req)
 { {
    _make_response(req, async_results::CIM_NAK);      MessageType type = req->getType();
       if (type == ASYNC_IOCTL)
       {
           handle_AsyncIoctl(static_cast<AsyncIoctl *>(req));
 } }
 void MessageQueueService::handle_CimServicePause(CimServicePause *req)      else if (type == ASYNC_CIMSERVICE_START)
 { {
    _make_response(req, async_results::CIM_NAK);          handle_CimServiceStart(static_cast<CimServiceStart *>(req));
 } }
 void MessageQueueService::handle_CimServiceResume(CimServiceResume *req)      else if (type == ASYNC_CIMSERVICE_STOP)
 { {
    _make_response(req, async_results::CIM_NAK);          handle_CimServiceStop(static_cast<CimServiceStop *>(req));
 } }
       else
 void MessageQueueService::handle_AsyncOperationStart(AsyncOperationStart *req)  
 { {
           // we don't handle this request message
    _make_response(req, async_results::CIM_NAK);    _make_response(req, async_results::CIM_NAK);
   
 } }
   
 void MessageQueueService::handle_AsyncOperationResult(AsyncOperationResult *req)  
 {  
    ;  
 } }
  
 AsyncOpNode *MessageQueueService::get_op(void)  Boolean MessageQueueService::_enqueueResponse(
       Message* request,
       Message* response)
 { {
    AsyncOpNode *op = new AsyncOpNode();      PEG_METHOD_ENTER(TRC_MESSAGEQUEUESERVICE,
           "MessageQueueService::_enqueueResponse");
  
    op->write_state(ASYNC_OPSTATE_UNKNOWN);      if (request->getMask() & MessageMask::ha_async)
    op->write_flags(ASYNC_OPFLAGS_SINGLE | ASYNC_OPFLAGS_NORMAL );      {
           if (response->getMask() & MessageMask::ha_async)
           {
               _completeAsyncResponse(
                   static_cast<AsyncRequest *>(request),
                   static_cast<AsyncReply *>(response));
  
    return op;              PEG_METHOD_EXIT();
               return true;
           }
 } }
  
 void MessageQueueService::return_op(AsyncOpNode *op)      AsyncRequest* asyncRequest =
           static_cast<AsyncRequest*>(request->get_async());
   
       if (asyncRequest != 0)
 { {
    PEGASUS_ASSERT(op->read_state() & ASYNC_OPSTATE_RELEASED );          PEGASUS_ASSERT(asyncRequest->getMask() &
    delete op;              (MessageMask::ha_async | MessageMask::ha_request));
 }  
  
           AsyncOpNode* op = asyncRequest->op;
  
           // the legacy request is going to be deleted by its handler
           // remove it from the op node
  
 AsyncReply *MessageQueueService::SendWait(AsyncRequest *request)          static_cast<AsyncLegacyOperationStart *>(asyncRequest)->get_action();
 {  
    if ( request == 0 )  
       return 0 ;  
    PEGASUS_ASSERT( request->op != 0 );  
  
    request->block = true;          AsyncLegacyOperationResult *async_result =
    request->op->_state &= ~ASYNC_OPSTATE_COMPLETE;              new AsyncLegacyOperationResult(
    request->op->put_response(0);                  op,
                   response);
           _completeAsyncResponse(
               asyncRequest,
               async_result);
  
    // first link it on our pending list          PEG_METHOD_EXIT();
    _pending.insert_last_wait(request->op);          return true;
       }
  
    // now see if the meta dispatcher will take it      // ensure that the destination queue is in response->dest
       PEG_METHOD_EXIT();
       return SendForget(response);
   }
  
    if (true == _meta_dispatcher->route_async(request->op))  void MessageQueueService::_make_response(Message* req, Uint32 code)
    {    {
       request->op->_client_sem.wait();      cimom::_make_response(req, code);
    }    }
  
    return static_cast<AsyncReply *>(request->op->get_response());  
  
 }  void MessageQueueService::_completeAsyncResponse(
       AsyncRequest* request,
       AsyncReply* reply)
   {
       PEG_METHOD_ENTER(TRC_MESSAGEQUEUESERVICE,
           "MessageQueueService::_completeAsyncResponse");
   
       cimom::_completeAsyncResponse(request, reply);
  
       PEG_METHOD_EXIT();
   }
  
 Boolean MessageQueueService::register_service(String name,  
                                               Uint32 capabilities,  
                                               Uint32 mask)  
  
   void MessageQueueService::_complete_op_node(
       AsyncOpNode* op)
 { {
    AsyncOpNode *op = _meta_dispatcher->get_cached_op();      cimom::_complete_op_node(op);
   }
  
    op->_state |= ASYNC_OPSTATE_UNKNOWN;  
    op->_flags |= ASYNC_OPFLAGS_SINGLE | ASYNC_OPFLAGS_NORMAL;  
  
    RegisterCimService *msg = new RegisterCimService(get_next_xid(),  Boolean MessageQueueService::accept_async(AsyncOpNode* op)
                                                     op,  
                                                     true,  
                                                     name,  
                                                     capabilities,  
                                                     mask,  
                                                     _queueId);  
    Boolean registered = false;  
    AsyncMessage *reply = SendWait( msg );  
   
    if ( reply != 0 )  
    {    {
       if(reply->getMask() & message_mask:: ha_async)      if (_incoming_queue_shutdown.get() > 0)
           return false;
       if (_polling_thread == NULL)
       {       {
          if(reply->getMask() & message_mask::ha_reply)          _polling_thread = new Thread(
               polling_routine,
               reinterpret_cast<void *>(_get_polling_list()),
               false);
           ThreadStatus tr = PEGASUS_THREAD_OK;
           while ( (tr =_polling_thread->run()) != PEGASUS_THREAD_OK)
          {          {
             if((static_cast<AsyncReply *>(reply))->result == async_results::OK)              if (tr == PEGASUS_THREAD_INSUFFICIENT_RESOURCES)
                registered = true;                  Threads::yield();
               else
                   throw Exception(MessageLoaderParms(
                       "Common.MessageQueueService.NOT_ENOUGH_THREAD",
                       "Could not allocate thread for the polling thread."));
          }          }
       }       }
       if (_die.get() == 0)
       delete reply;      {
           if (_incoming.enqueue(op))
           {
               _polling_sem.signal();
               return true;
    }    }
    return registered;      }
       return false;
 } }
  
 Boolean MessageQueueService::update_service(Uint32 capabilities, Uint32 mask)  void MessageQueueService::handle_AsyncIoctl(AsyncIoctl* req)
 { {
       switch (req->ctl)
       {
           case AsyncIoctl::IO_CLOSE:
           {
               MessageQueueService *service =
                   static_cast<MessageQueueService *>(req->op->_service_ptr);
  
    AsyncOpNode *op = _meta_dispatcher->get_cached_op();  #ifdef MESSAGEQUEUESERVICE_DEBUG
    op->_state |= ASYNC_OPSTATE_UNKNOWN;              PEGASUS_STD(cout) << service->getQueueName() <<
    op->_flags |= ASYNC_OPFLAGS_SINGLE | ASYNC_OPFLAGS_NORMAL;                  " Received AsyncIoctl::IO_CLOSE " << PEGASUS_STD(endl);
   #endif
    UpdateCimService *msg = new UpdateCimService(get_next_xid(),  
                                                 op,              // respond to this message. this is fire and forget, so we
                                                 true,              // don't need to delete anything.
                                                 _queueId,              // this takes care of two problems that were being found
                                                 _capabilities,              // << Thu Oct  9 10:52:48 2003 mdd >>
                                                 _mask);              _make_response(req, async_results::OK);
    Boolean registered = false;              // ensure we do not accept any further messages
   
  
    AsyncMessage *reply = SendWait(msg);              // ensure we don't recurse on IO_CLOSE
    if (reply)              if (_incoming_queue_shutdown.get() > 0)
    {                  break;
       if(reply->getMask() & message_mask:: ha_async)  
               // set the closing flag
               service->_incoming_queue_shutdown = 1;
               // empty out the queue
               while (1)
       {       {
          if(reply->getMask() & message_mask::ha_reply)                  AsyncOpNode* operation = 0;
                   try
          {          {
             if(static_cast<AsyncReply *>(reply)->result == async_results::OK)                      operation = service->_incoming.dequeue();
                registered = true;  
          }          }
                   catch (...)
                   {
                       break;
       }       }
       delete reply;                  if (operation)
                   {
                       operation->_service_ptr = service;
                       service->_handle_incoming_operation(operation);
    }    }
    return registered;                  else
                       break;
               } // message processing loop
   
               // shutdown the AsyncQueue
               service->_incoming.close();
               return;
 } }
  
           default:
               _make_response(req, async_results::CIM_NAK);
       }
   }
  
 Boolean MessageQueueService::deregister_service(void)  void MessageQueueService::handle_CimServiceStart(CimServiceStart* req)
 { {
   #ifdef MESSAGEQUEUESERVICE_DEBUG
       PEGASUS_STD(cout) << getQueueName() << "received START" <<
           PEGASUS_STD(endl);
   #endif
       PEGASUS_ASSERT(!_isRunning);
       _isRunning = true;
       _make_response(req, async_results::OK);
   }
  
 //   _meta_dispatcher->deregister_module(_queueId);  void MessageQueueService::handle_CimServiceStop(CimServiceStop* req)
 //   return true;  {
   #ifdef MESSAGEQUEUESERVICE_DEBUG
       PEGASUS_STD(cout) << getQueueName() << "received STOP" << PEGASUS_STD(endl);
    AsyncOpNode *op = _meta_dispatcher->get_cached_op();  #endif
    op->_state |= ASYNC_OPSTATE_UNKNOWN;      PEGASUS_ASSERT(_isRunning);
    op->_flags |= ASYNC_OPFLAGS_SINGLE | ASYNC_OPFLAGS_NORMAL;      _isRunning = false;
       _make_response(req, async_results::CIM_SERVICE_STOPPED);
   }
  
    DeRegisterCimService *msg = new DeRegisterCimService(get_next_xid(),  AsyncOpNode* MessageQueueService::get_op()
                                                         op,  {
                                                         true,     AsyncOpNode* op = new AsyncOpNode();
                                                         _queueId);  
    Boolean deregistered = false;  
  
      op->_state = ASYNC_OPSTATE_UNKNOWN;
      op->_flags = ASYNC_OPFLAGS_UNKNOWN;
  
      return op;
   }
  
    return _meta_dispatcher->accept_async(static_cast<Message *>(msg));  void MessageQueueService::return_op(AsyncOpNode* op)
   {
       delete op;
 } }
  
  
 void MessageQueueService::find_services(String name,  Boolean MessageQueueService::SendAsync(
                                         Uint32 capabilities,      AsyncOpNode* op,
                                         Uint32 mask,      Uint32 destination,
                                         Array<Uint32> *results)      void (*callback)(AsyncOpNode*, MessageQueue*, void*),
       MessageQueue* callback_response_q,
       void* callback_ptr)
 { {
       return _sendAsync(
           op,
           destination,
           callback,
           callback_response_q,
           callback_ptr,
           ASYNC_OPFLAGS_CALLBACK);
  
    if( results == 0 )  }
       throw NullPointer();  
  
    AsyncOpNode *op = get_op();  Boolean MessageQueueService::_sendAsync(
    results->clear();      AsyncOpNode* op,
       Uint32 destination,
       void (*callback)(AsyncOpNode*, MessageQueue*, void*),
       MessageQueue* callback_response_q,
       void* callback_ptr,
       Uint32 flags)
   {
       PEGASUS_ASSERT(op != 0 && callback != 0);
   
       // destination of this message
       op->_op_dest = MessageQueue::lookup(destination);
       if (op->_op_dest == 0)
       {
           return false;
       }
       op->_flags = flags;
       // initialize the callback data
       // callback function to be executed by recpt. of response
       op->_async_callback = callback;
       // the op node
       op->_callback_node = op;
       // the queue that will receive the response
       op->_callback_response_q = callback_response_q;
       // user data for callback
       op->_callback_ptr = callback_ptr;
       // I am the originator of this request
       op->_callback_request_q = this;
  
    FindServiceQueue *req =      return  _meta_dispatcher->route_async(op);
       new FindServiceQueue(get_next_xid(),  }
                            op,  
                            _queueId,  
                            true,  
                            name,  
                            capabilities,  
                            mask);  
  
    AsyncMessage *reply = SendWait(req);  Boolean MessageQueueService::SendForget(Message* msg)
    if(reply)  
    {    {
       if( reply->getMask() & message_mask::ha_async)      AsyncOpNode* op = 0;
       Uint32 mask = msg->getMask();
   
       if (mask & MessageMask::ha_async)
       {       {
          if(reply->getMask() & message_mask::ha_reply)          op = (static_cast<AsyncMessage *>(msg))->op;
       }
   
       if (op == 0)
          {          {
             if(reply->getType() == async_messages::FIND_SERVICE_Q_RESULT)          op = get_op();
           op->_request.reset(msg);
           if (mask & MessageMask::ha_async)
             {             {
                if( (static_cast<FindServiceQueueResult *>(reply))->result == async_results::OK )              (static_cast<AsyncMessage *>(msg))->op = op;
                   *results = (static_cast<FindServiceQueueResult *>(reply))->qids;  
             }             }
          }          }
       }  
       delete reply;  
    }  
    return ;  
 }  
  
 void MessageQueueService::enumerate_service(Uint32 queue, message_module *result)      PEGASUS_ASSERT(op->_flags == ASYNC_OPFLAGS_UNKNOWN);
       PEGASUS_ASSERT(op->_state == ASYNC_OPSTATE_UNKNOWN);
       op->_op_dest = MessageQueue::lookup(msg->dest);
       if (op->_op_dest == 0)
 { {
    if(result == 0)          return_op(op);
       throw NullPointer();          return false;
       }
  
    AsyncOpNode *op = get_op();      op->_flags = ASYNC_OPFLAGS_FIRE_AND_FORGET;
  
    EnumerateService *req      // now see if the meta dispatcher will take it
       = new EnumerateService(get_next_xid(),      return  _meta_dispatcher->route_async(op);
                              op,  }
                              _queueId,  
                              true,  
                              queue);  
  
    AsyncMessage *reply = SendWait(req);  
  
    if (reply)  AsyncReply *MessageQueueService::SendWait(AsyncRequest* request)
    {    {
       Boolean found = false;      if (request == 0)
           return 0;
  
       if( reply->getMask() & message_mask::ha_async)      Boolean destroy_op = false;
       {  
          if(reply->getMask() & message_mask::ha_reply)      if (request->op == 0)
          {  
             if(reply->getType() == async_messages::ENUMERATE_SERVICE_RESULT)  
             {  
                if( (static_cast<EnumerateServiceResponse *>(reply))->result == async_results::OK )  
                {  
                   if( found == false)  
                   {                   {
                      found = true;          request->op = get_op();
           request->op->_request.reset(request);
           destroy_op = true;
       }
   
       PEGASUS_ASSERT(request->op->_flags == ASYNC_OPFLAGS_UNKNOWN);
       PEGASUS_ASSERT(request->op->_state == ASYNC_OPSTATE_UNKNOWN);
   
       request->block = false;
       _sendAsync(
           request->op,
           request->dest,
           _sendwait_callback,
           this,
           (void *)0,
           ASYNC_OPFLAGS_PSEUDO_CALLBACK);
  
                      result->put_name( (static_cast<EnumerateServiceResponse *>(reply))->name);      request->op->_client_sem.wait();
                      result->put_capabilities((static_cast<EnumerateServiceResponse *>(reply))->capabilities);  
                      result->put_mask((static_cast<EnumerateServiceResponse *>(reply))->mask);      AsyncReply* rpl = static_cast<AsyncReply *>(request->op->removeResponse());
                      result->put_queue((static_cast<EnumerateServiceResponse *>(reply))->qid);      rpl->op = 0;
                   }  
                }      if (destroy_op == true)
             }      {
          }          request->op->_request.release();
           return_op(request->op);
           request->op = 0;
       }       }
       delete reply;      return rpl;
    }    }
    return;  
   Uint32 MessageQueueService::find_service_qid(const String &name)
   {
       MessageQueue *queue = MessageQueue::lookup((const char*)name.getCString());
       PEGASUS_ASSERT(queue);
       return queue->getQueueId();
 } }
  
 Uint32 MessageQueueService::get_next_xid(void)  MessageQueueService::PollingList* MessageQueueService::_get_polling_list()
 { {
    _xid++;      _polling_list_mutex.lock();
    return _xid.value();  
       if (!_polling_list)
           _polling_list = new PollingList;
   
       _polling_list_mutex.unlock();
   
       return _polling_list;
 } }
  
 PEGASUS_NAMESPACE_END PEGASUS_NAMESPACE_END


Legend:
Removed from v.1.4  
changed lines
  Added in v.1.151

No CVS admin address has been configured
Powered by
ViewCVS 0.9.2