version 1.124, 2006/08/09 21:12:42
|
version 1.138, 2008/06/19 17:57:01
|
|
|
// | // |
//============================================================================== | //============================================================================== |
// | // |
// Author: Mike Day (mdday@us.ibm.com) |
|
// |
|
// Modified By: |
|
// Amit K Arora, IBM (amita@in.ibm.com) for Bug#1090,#2657 |
|
// Josephine Eskaline Joyce, IBM (jojustin@in.ibm.com) for Bug#3259 |
|
// Jim Wunderlich (Jim_Wunderlich@prodigy.net) |
|
// |
|
//%///////////////////////////////////////////////////////////////////////////// | //%///////////////////////////////////////////////////////////////////////////// |
| |
// #include <iostream.h> |
|
#include "MessageQueueService.h" | #include "MessageQueueService.h" |
#include <Pegasus/Common/Tracer.h> | #include <Pegasus/Common/Tracer.h> |
#include <Pegasus/Common/MessageLoader.h> //l10n |
#include <Pegasus/Common/MessageLoader.h> |
| |
PEGASUS_NAMESPACE_BEGIN | PEGASUS_NAMESPACE_BEGIN |
| |
|
|
| |
Uint32 max_threads_per_svc_queue; | Uint32 max_threads_per_svc_queue; |
| |
ThreadReturnType PEGASUS_THREAD_CDECL |
ThreadReturnType PEGASUS_THREAD_CDECL MessageQueueService::polling_routine( |
MessageQueueService::kill_idle_threads(void *parm) |
void* parm) |
{ |
|
|
|
static struct timeval now, last = {0,0}; |
|
Time::gettimeofday(&now); |
|
int dead_threads = 0; |
|
|
|
if (now.tv_sec - last.tv_sec > 120) |
|
{ |
|
Time::gettimeofday(&last); |
|
try |
|
{ |
|
dead_threads = MessageQueueService::_thread_pool->cleanupIdleThreads(); |
|
} |
|
catch(...) |
|
{ |
|
|
|
} |
|
} |
|
|
|
#ifdef PEGASUS_POINTER_64BIT |
|
return (ThreadReturnType)(Uint64)dead_threads; |
|
#elif PEGASUS_PLATFORM_AIX_RS_IBMCXX |
|
return (ThreadReturnType)(unsigned long)dead_threads; |
|
#else |
|
return (ThreadReturnType)(Uint32)dead_threads; |
|
#endif |
|
} |
|
|
|
ThreadReturnType PEGASUS_THREAD_CDECL MessageQueueService::polling_routine(void *parm) |
|
{ | { |
Thread *myself = reinterpret_cast<Thread *>(parm); | Thread *myself = reinterpret_cast<Thread *>(parm); |
List<MessageQueueService, Mutex> *list = | List<MessageQueueService, Mutex> *list = |
|
|
} | } |
| |
// The polling_routine thread must hold the lock on the | // The polling_routine thread must hold the lock on the |
// _polling_thread list while processing incoming messages. |
// _polling_list while processing incoming messages. |
// This lock is used to give this thread ownership of | // This lock is used to give this thread ownership of |
// services on the _polling_routine list. | // services on the _polling_routine list. |
| |
|
|
if (rtn != PEGASUS_THREAD_OK ) | if (rtn != PEGASUS_THREAD_OK ) |
{ | { |
service->_threads--; | service->_threads--; |
Logger::put(Logger::STANDARD_LOG, System::CIMSERVER, Logger::TRACE, |
PEG_TRACE((TRC_MESSAGEQUEUESERVICE, Tracer::LEVEL1, |
"Not enough threads to process this request. Skipping."); |
"Could not allocate thread for %s. Queue has %d " |
|
"messages waiting and %d threads servicing." |
Tracer::trace(TRC_MESSAGEQUEUESERVICE, Tracer::LEVEL2, |
|
"Could not allocate thread for %s. " \ |
|
"Queue has %d messages waiting and %d threads servicing." \ |
|
"Skipping the service for right now. ", | "Skipping the service for right now. ", |
service->getQueueName(), | service->getQueueName(), |
service->_incoming.count(), | service->_incoming.count(), |
service->_threads.get()); |
service->_threads.get())); |
| |
Threads::yield(); | Threads::yield(); |
service = NULL; | service = NULL; |
|
|
} | } |
} | } |
list->unlock(); | list->unlock(); |
|
|
if (_check_idle_flag.get() != 0) |
|
{ |
|
_check_idle_flag = 0; |
|
// try to do idle thread clean up processing when system is not busy |
|
// if system is busy there may not be a thread available to allocate |
|
// so nothing will be done and that is OK. |
|
|
|
if ( _thread_pool->allocate_and_awaken(service, kill_idle_threads, &_polling_sem) != PEGASUS_THREAD_OK) |
|
{ |
|
Logger::put(Logger::STANDARD_LOG, System::CIMSERVER, Logger::TRACE, |
|
"Not enough threads to kill idle threads. What an irony."); |
|
|
|
Tracer::trace(TRC_MESSAGEQUEUESERVICE, Tracer::LEVEL2, |
|
"Could not allocate thread to kill idle threads." \ |
|
"Skipping. "); |
|
} |
|
|
|
|
|
} |
|
} | } |
myself->exit_self( (ThreadReturnType) 1 ); |
return ThreadReturnType(0); |
return(0); |
|
} | } |
| |
| |
Semaphore MessageQueueService::_polling_sem(0); | Semaphore MessageQueueService::_polling_sem(0); |
AtomicInt MessageQueueService::_stop_polling(0); | AtomicInt MessageQueueService::_stop_polling(0); |
AtomicInt MessageQueueService::_check_idle_flag(0); |
|
| |
| |
MessageQueueService::MessageQueueService( | MessageQueueService::MessageQueueService( |
|
|
_incoming(), | _incoming(), |
_incoming_queue_shutdown(0) | _incoming_queue_shutdown(0) |
{ | { |
|
|
_capabilities = (capabilities | module_capabilities::async); | _capabilities = (capabilities | module_capabilities::async); |
| |
_default_op_timeout.tv_sec = 30; | _default_op_timeout.tv_sec = 30; |
|
|
max_threads_per_svc_queue = MAX_THREADS_PER_SVC_QUEUE_LIMIT; | max_threads_per_svc_queue = MAX_THREADS_PER_SVC_QUEUE_LIMIT; |
} | } |
| |
Tracer::trace(TRC_MESSAGEQUEUESERVICE, Tracer::LEVEL2, |
PEG_TRACE((TRC_MESSAGEQUEUESERVICE, Tracer::LEVEL3, |
"max_threads_per_svc_queue set to %u.", max_threads_per_svc_queue); |
"max_threads_per_svc_queue set to %u.", max_threads_per_svc_queue)); |
| |
AutoMutex autoMut(_meta_dispatcher_mutex); | AutoMutex autoMut(_meta_dispatcher_mutex); |
| |
|
|
_stop_polling = 0; | _stop_polling = 0; |
PEGASUS_ASSERT(_service_count.get() == 0); | PEGASUS_ASSERT(_service_count.get() == 0); |
_meta_dispatcher = new cimom(); | _meta_dispatcher = new cimom(); |
if (_meta_dispatcher == NULL) |
|
{ |
|
throw NullPointer(); |
|
} |
|
// _thread_pool = new ThreadPool(initial_cnt, "MessageQueueService", | // _thread_pool = new ThreadPool(initial_cnt, "MessageQueueService", |
// minimum_cnt, maximum_cnt, deallocateWait); | // minimum_cnt, maximum_cnt, deallocateWait); |
// | // |
|
|
| |
if (false == register_service(name, _capabilities, _mask)) | if (false == register_service(name, _capabilities, _mask)) |
{ | { |
//l10n |
MessageLoaderParms parms( |
//throw BindFailedException("MessageQueueService Base Unable to register with Meta Dispatcher"); |
"Common.MessageQueueService.UNABLE_TO_REGISTER", |
MessageLoaderParms parms("Common.MessageQueueService.UNABLE_TO_REGISTER", |
"CIM base message queue service is unable to register with the " |
"MessageQueueService Base Unable to register with Meta Dispatcher"); |
"CIMOM dispatcher."); |
|
|
throw BindFailedException(parms); | throw BindFailedException(parms); |
} | } |
| |
_get_polling_list()->insert_back(this); | _get_polling_list()->insert_back(this); |
|
|
} | } |
| |
| |
|
|
| |
_stop_polling++; | _stop_polling++; |
_polling_sem.signal(); | _polling_sem.signal(); |
if (_polling_thread) { |
if (_polling_thread) |
|
{ |
_polling_thread->join(); | _polling_thread->join(); |
delete _polling_thread; | delete _polling_thread; |
_polling_thread = 0; | _polling_thread = 0; |
|
|
// Clean up in case there are extra stuff on the queue. | // Clean up in case there are extra stuff on the queue. |
while (_incoming.count()) | while (_incoming.count()) |
{ | { |
try { |
try |
|
{ |
delete _incoming.dequeue(); | delete _incoming.dequeue(); |
} catch (const ListClosed &e) |
} |
|
catch (const ListClosed&) |
{ | { |
// If the list is closed, there is nothing we can do. | // If the list is closed, there is nothing we can do. |
break; | break; |
|
|
| |
msg->op->_op_dest = this; | msg->op->_op_dest = this; |
msg->op->_request.reset(msg); | msg->op->_request.reset(msg); |
try { |
try |
|
{ |
_incoming.enqueue_wait(msg->op); | _incoming.enqueue_wait(msg->op); |
_polling_sem.signal(); | _polling_sem.signal(); |
} catch (const ListClosed &) |
} |
|
catch (const ListClosed&) |
{ | { |
// This means the queue has already been shut-down (happens when there | // This means the queue has already been shut-down (happens when there |
// are two AsyncIoctrl::IO_CLOSE messages generated and one got first | // are two AsyncIoctrl::IO_CLOSE messages generated and one got first |
|
|
PEGASUS_ASSERT(service != 0); | PEGASUS_ASSERT(service != 0); |
try | try |
{ | { |
|
|
if (service->_die.get() != 0) | if (service->_die.get() != 0) |
{ | { |
service->_threads--; | service->_threads--; |
return (0); |
return 0; |
} | } |
// pull messages off the incoming queue and dispatch them. then | // pull messages off the incoming queue and dispatch them. then |
// check pending messages that are non-blocking | // check pending messages that are non-blocking |
|
|
catch (ListClosed &) | catch (ListClosed &) |
{ | { |
// ATTN: This appears to be a common loop exit path. | // ATTN: This appears to be a common loop exit path. |
//PEG_TRACE_STRING(TRC_DISCARDED_DATA, Tracer::LEVEL2, |
//PEG_TRACE_CSTRING(TRC_DISCARDED_DATA, Tracer::LEVEL2, |
// "Caught ListClosed exception. Exiting _req_proc."); | // "Caught ListClosed exception. Exiting _req_proc."); |
break; | break; |
} | } |
|
|
} | } |
catch (const Exception& e) | catch (const Exception& e) |
{ | { |
PEG_TRACE_STRING(TRC_DISCARDED_DATA, Tracer::LEVEL2, |
PEG_TRACE_STRING(TRC_DISCARDED_DATA, Tracer::LEVEL1, |
String("Caught exception: \"") + e.getMessage() + | String("Caught exception: \"") + e.getMessage() + |
"\". Exiting _req_proc."); | "\". Exiting _req_proc."); |
} | } |
catch (...) | catch (...) |
{ | { |
PEG_TRACE_STRING(TRC_DISCARDED_DATA, Tracer::LEVEL2, |
PEG_TRACE_CSTRING(TRC_DISCARDED_DATA, Tracer::LEVEL1, |
"Caught unrecognized exception. Exiting _req_proc."); | "Caught unrecognized exception. Exiting _req_proc."); |
} | } |
service->_threads--; | service->_threads--; |
return(0); |
return 0; |
} | } |
| |
| |
|
|
{ | { |
if (op->_flags & ASYNC_OPFLAGS_SAFE_CALLBACK) | if (op->_flags & ASYNC_OPFLAGS_SAFE_CALLBACK) |
{ | { |
|
|
Message *msg = op->removeRequest(); | Message *msg = op->removeRequest(); |
if (msg && (msg->getMask() & message_mask::ha_async)) |
if (msg && (msg->getMask() & MessageMask::ha_async)) |
{ | { |
if (msg->getType() == async_messages::ASYNC_LEGACY_OP_START) |
if (msg->getType() == ASYNC_ASYNC_LEGACY_OP_START) |
{ | { |
AsyncLegacyOperationStart *wrapper = | AsyncLegacyOperationStart *wrapper = |
static_cast<AsyncLegacyOperationStart *>(msg); | static_cast<AsyncLegacyOperationStart *>(msg); |
msg = wrapper->get_action(); | msg = wrapper->get_action(); |
delete wrapper; | delete wrapper; |
} | } |
else if (msg->getType() == async_messages::ASYNC_MODULE_OP_START) |
else if (msg->getType() == ASYNC_ASYNC_MODULE_OP_START) |
{ | { |
AsyncModuleOperationStart *wrapper = | AsyncModuleOperationStart *wrapper = |
static_cast<AsyncModuleOperationStart *>(msg); | static_cast<AsyncModuleOperationStart *>(msg); |
msg = wrapper->get_action(); | msg = wrapper->get_action(); |
delete wrapper; | delete wrapper; |
} | } |
else if (msg->getType() == async_messages::ASYNC_OP_START) |
else if (msg->getType() == ASYNC_ASYNC_OP_START) |
{ | { |
AsyncOperationStart *wrapper = | AsyncOperationStart *wrapper = |
static_cast<AsyncOperationStart *>(msg); | static_cast<AsyncOperationStart *>(msg); |
|
|
} | } |
| |
msg = op->removeResponse(); | msg = op->removeResponse(); |
if (msg && (msg->getMask() & message_mask::ha_async)) |
if (msg && (msg->getMask() & MessageMask::ha_async)) |
{ | { |
if (msg->getType() == async_messages::ASYNC_LEGACY_OP_RESULT) |
if (msg->getType() == ASYNC_ASYNC_LEGACY_OP_RESULT) |
{ | { |
AsyncLegacyOperationResult *wrapper = | AsyncLegacyOperationResult *wrapper = |
static_cast<AsyncLegacyOperationResult *>(msg); | static_cast<AsyncLegacyOperationResult *>(msg); |
msg = wrapper->get_result(); | msg = wrapper->get_result(); |
delete wrapper; | delete wrapper; |
} | } |
else if (msg->getType() == async_messages::ASYNC_MODULE_OP_RESULT) |
else if (msg->getType() == ASYNC_ASYNC_MODULE_OP_RESULT) |
{ | { |
AsyncModuleOperationResult *wrapper = | AsyncModuleOperationResult *wrapper = |
static_cast<AsyncModuleOperationResult *>(msg); | static_cast<AsyncModuleOperationResult *>(msg); |
|
|
// note that _callback_node may be different from op | // note that _callback_node may be different from op |
// op->_callback_response_q is a "this" pointer we can use for | // op->_callback_response_q is a "this" pointer we can use for |
// static callback methods | // static callback methods |
op->_async_callback(op->_callback_node, op->_callback_response_q, op->_callback_ptr); |
op->_async_callback( |
|
op->_callback_node, op->_callback_response_q, op->_callback_ptr); |
} | } |
} | } |
| |
|
|
// messages become async messages. | // messages become async messages. |
| |
// divert legacy messages to handleEnqueue | // divert legacy messages to handleEnqueue |
if ((rq != 0) && (!(rq->getMask() & message_mask::ha_async))) |
if ((rq != 0) && (!(rq->getMask() & MessageMask::ha_async))) |
{ | { |
operation->_request.release(); | operation->_request.release(); |
operation->unlock(); | operation->unlock(); |
|
|
{ | { |
req->op->processing(); | req->op->processing(); |
| |
Uint32 type = req->getType(); |
MessageType type = req->getType(); |
if (type == async_messages::HEARTBEAT) |
if (type == ASYNC_HEARTBEAT) |
handle_heartbeat_request(req); | handle_heartbeat_request(req); |
else if (type == async_messages::IOCTL) |
else if (type == ASYNC_IOCTL) |
handle_AsyncIoctl(static_cast<AsyncIoctl *>(req)); | handle_AsyncIoctl(static_cast<AsyncIoctl *>(req)); |
else if (type == async_messages::CIMSERVICE_START) |
else if (type == ASYNC_CIMSERVICE_START) |
handle_CimServiceStart(static_cast<CimServiceStart *>(req)); | handle_CimServiceStart(static_cast<CimServiceStart *>(req)); |
else if (type == async_messages::CIMSERVICE_STOP) |
else if (type == ASYNC_CIMSERVICE_STOP) |
handle_CimServiceStop(static_cast<CimServiceStop *>(req)); | handle_CimServiceStop(static_cast<CimServiceStop *>(req)); |
else if (type == async_messages::CIMSERVICE_PAUSE) |
else if (type == ASYNC_CIMSERVICE_PAUSE) |
handle_CimServicePause(static_cast<CimServicePause *>(req)); | handle_CimServicePause(static_cast<CimServicePause *>(req)); |
else if (type == async_messages::CIMSERVICE_RESUME) |
else if (type == ASYNC_CIMSERVICE_RESUME) |
handle_CimServiceResume(static_cast<CimServiceResume *>(req)); | handle_CimServiceResume(static_cast<CimServiceResume *>(req)); |
else if (type == async_messages::ASYNC_OP_START) |
else if (type == ASYNC_ASYNC_OP_START) |
handle_AsyncOperationStart(static_cast<AsyncOperationStart *>(req)); | handle_AsyncOperationStart(static_cast<AsyncOperationStart *>(req)); |
else | else |
{ | { |
|
|
Message* request, | Message* request, |
Message* response) | Message* response) |
{ | { |
|
|
STAT_COPYDISPATCHER |
|
|
|
PEG_METHOD_ENTER(TRC_MESSAGEQUEUESERVICE, | PEG_METHOD_ENTER(TRC_MESSAGEQUEUESERVICE, |
"MessageQueueService::_enqueueResponse"); | "MessageQueueService::_enqueueResponse"); |
| |
if (request->getMask() & message_mask::ha_async) |
if (request->getMask() & MessageMask::ha_async) |
{ | { |
if (response->getMask() & message_mask::ha_async) |
if (response->getMask() & MessageMask::ha_async) |
{ | { |
_completeAsyncResponse(static_cast<AsyncRequest *>(request), |
_completeAsyncResponse( |
|
static_cast<AsyncRequest *>(request), |
static_cast<AsyncReply *>(response), | static_cast<AsyncReply *>(response), |
ASYNC_OPSTATE_COMPLETE, 0); | ASYNC_OPSTATE_COMPLETE, 0); |
PEG_METHOD_EXIT(); | PEG_METHOD_EXIT(); |
|
|
} | } |
} | } |
| |
if (request->_async != 0) |
AsyncRequest* asyncRequest = |
|
static_cast<AsyncRequest*>(request->get_async()); |
|
|
|
if (asyncRequest != 0) |
{ | { |
Uint32 mask = request->_async->getMask(); |
PEGASUS_ASSERT(asyncRequest->getMask() & |
PEGASUS_ASSERT(mask & (message_mask::ha_async | message_mask::ha_request)); |
(MessageMask::ha_async | MessageMask::ha_request)); |
|
|
|
AsyncOpNode* op = asyncRequest->op; |
| |
AsyncRequest *async = static_cast<AsyncRequest *>(request->_async); |
|
AsyncOpNode *op = async->op; |
|
request->_async = 0; |
|
// the legacy request is going to be deleted by its handler | // the legacy request is going to be deleted by its handler |
// remove it from the op node | // remove it from the op node |
| |
static_cast<AsyncLegacyOperationStart *>(async)->get_action(); |
static_cast<AsyncLegacyOperationStart *>(asyncRequest)->get_action(); |
| |
AsyncLegacyOperationResult *async_result = | AsyncLegacyOperationResult *async_result = |
new AsyncLegacyOperationResult( | new AsyncLegacyOperationResult( |
op, | op, |
response); | response); |
_completeAsyncResponse( | _completeAsyncResponse( |
async, |
asyncRequest, |
async_result, | async_result, |
ASYNC_OPSTATE_COMPLETE, | ASYNC_OPSTATE_COMPLETE, |
0); | 0); |
|
|
// ensure that the destination queue is in response->dest | // ensure that the destination queue is in response->dest |
PEG_METHOD_EXIT(); | PEG_METHOD_EXIT(); |
return SendForget(response); | return SendForget(response); |
|
|
} | } |
| |
void MessageQueueService::_make_response(Message *req, Uint32 code) | void MessageQueueService::_make_response(Message *req, Uint32 code) |
|
|
if (tr == PEGASUS_THREAD_INSUFFICIENT_RESOURCES) | if (tr == PEGASUS_THREAD_INSUFFICIENT_RESOURCES) |
Threads::yield(); | Threads::yield(); |
else | else |
throw Exception(MessageLoaderParms("Common.MessageQueueService.NOT_ENOUGH_THREAD", |
throw Exception(MessageLoaderParms( |
|
"Common.MessageQueueService.NOT_ENOUGH_THREAD", |
"Could not allocate thread for the polling thread.")); | "Could not allocate thread for the polling thread.")); |
} | } |
} | } |
|
|
// default action is to echo a heartbeat response | // default action is to echo a heartbeat response |
| |
AsyncReply *reply = new AsyncReply( | AsyncReply *reply = new AsyncReply( |
async_messages::HEARTBEAT, |
ASYNC_HEARTBEAT, |
0, | 0, |
req->op, | req->op, |
async_results::OK, | async_results::OK, |
|
|
{ | { |
case AsyncIoctl::IO_CLOSE: | case AsyncIoctl::IO_CLOSE: |
{ | { |
MessageQueueService *service = static_cast<MessageQueueService *>(req->op->_service_ptr); |
MessageQueueService *service = |
|
static_cast<MessageQueueService *>(req->op->_service_ptr); |
| |
#ifdef MESSAGEQUEUESERVICE_DEBUG | #ifdef MESSAGEQUEUESERVICE_DEBUG |
PEGASUS_STD(cout) << service->getQueueName() << " Received AsyncIoctl::IO_CLOSE " << PEGASUS_STD(endl); |
PEGASUS_STD(cout) << service->getQueueName() << |
|
" Received AsyncIoctl::IO_CLOSE " << PEGASUS_STD(endl); |
#endif | #endif |
| |
// respond to this message. this is fire and forget, so we don't need to delete anything. |
// respond to this message. this is fire and forget, so we |
|
// don't need to delete anything. |
// this takes care of two problems that were being found | // this takes care of two problems that were being found |
// << Thu Oct 9 10:52:48 2003 mdd >> | // << Thu Oct 9 10:52:48 2003 mdd >> |
_make_response(req, async_results::OK); | _make_response(req, async_results::OK); |
|
|
| |
void MessageQueueService::handle_CimServiceStart(CimServiceStart *req) | void MessageQueueService::handle_CimServiceStart(CimServiceStart *req) |
{ | { |
|
|
#ifdef MESSAGEQUEUESERVICE_DEBUG | #ifdef MESSAGEQUEUESERVICE_DEBUG |
PEGASUS_STD(cout) << getQueueName() << "received START" << PEGASUS_STD(endl); |
PEGASUS_STD(cout) << getQueueName() << "received START" << |
|
PEGASUS_STD(endl); |
#endif | #endif |
| |
// clear the stoped bit and update | // clear the stoped bit and update |
|
|
_make_response(req, async_results::OK); | _make_response(req, async_results::OK); |
// now tell the meta dispatcher we are stopped | // now tell the meta dispatcher we are stopped |
update_service(_capabilities, _mask); | update_service(_capabilities, _mask); |
|
|
} | } |
|
|
void MessageQueueService::handle_CimServiceStop(CimServiceStop *req) | void MessageQueueService::handle_CimServiceStop(CimServiceStop *req) |
{ | { |
#ifdef MESSAGEQUEUESERVICE_DEBUG | #ifdef MESSAGEQUEUESERVICE_DEBUG |
|
|
| |
void MessageQueueService::handle_AsyncOperationResult(AsyncOperationResult *req) | void MessageQueueService::handle_AsyncOperationResult(AsyncOperationResult *req) |
{ | { |
; |
|
} | } |
| |
| |
void MessageQueueService::handle_AsyncLegacyOperationStart(AsyncLegacyOperationStart *req) |
void MessageQueueService::handle_AsyncLegacyOperationStart( |
|
AsyncLegacyOperationStart* req) |
{ | { |
// remove the legacy message from the request and enqueue it to its destination |
// remove the legacy message from the request and enqueue it to its |
|
// destination |
Uint32 result = async_results::CIM_NAK; | Uint32 result = async_results::CIM_NAK; |
| |
Message *legacy = req->_act; | Message *legacy = req->_act; |
|
|
{ | { |
if (queue->isAsync() == true) | if (queue->isAsync() == true) |
{ | { |
(static_cast<MessageQueueService *>(queue))->handleEnqueue(legacy); |
(static_cast<MessageQueueService *>(queue))->handleEnqueue( |
|
legacy); |
} | } |
else | else |
{ | { |
|
|
_make_response(req, result); | _make_response(req, result); |
} | } |
| |
void MessageQueueService::handle_AsyncLegacyOperationResult(AsyncLegacyOperationResult *rep) |
void MessageQueueService::handle_AsyncLegacyOperationResult( |
|
AsyncLegacyOperationResult* rep) |
{ | { |
; |
|
} | } |
| |
AsyncOpNode *MessageQueueService::get_op() | AsyncOpNode *MessageQueueService::get_op() |
|
|
} | } |
| |
| |
Boolean MessageQueueService::ForwardOp( |
|
AsyncOpNode *op, |
|
Uint32 destination) |
|
{ |
|
PEGASUS_ASSERT(op != 0); |
|
op->lock(); |
|
op->_op_dest = MessageQueue::lookup(destination); |
|
op->_flags |= (ASYNC_OPFLAGS_FIRE_AND_FORGET | ASYNC_OPFLAGS_FORWARD); |
|
op->_flags &= ~(ASYNC_OPFLAGS_CALLBACK); |
|
op->unlock(); |
|
if (op->_op_dest == 0) |
|
return false; |
|
|
|
return _meta_dispatcher->route_async(op); |
|
} |
|
|
|
|
|
Boolean MessageQueueService::SendAsync( | Boolean MessageQueueService::SendAsync( |
AsyncOpNode *op, | AsyncOpNode *op, |
Uint32 destination, | Uint32 destination, |
|
|
// get the queue handle for the destination | // get the queue handle for the destination |
| |
op->lock(); | op->lock(); |
op->_op_dest = MessageQueue::lookup(destination); // destination of this message |
// destination of this message |
|
op->_op_dest = MessageQueue::lookup(destination); |
op->_flags |= ASYNC_OPFLAGS_CALLBACK; | op->_flags |= ASYNC_OPFLAGS_CALLBACK; |
op->_flags &= ~(ASYNC_OPFLAGS_FIRE_AND_FORGET); | op->_flags &= ~(ASYNC_OPFLAGS_FIRE_AND_FORGET); |
// initialize the callback data | // initialize the callback data |
op->_async_callback = callback; // callback function to be executed by recpt. of response |
// callback function to be executed by recpt. of response |
op->_callback_node = op; // the op node |
op->_async_callback = callback; |
op->_callback_response_q = callback_response_q; // the queue that will receive the response |
// the op node |
op->_callback_ptr = callback_ptr; // user data for callback |
op->_callback_node = op; |
op->_callback_request_q = this; // I am the originator of this request |
// the queue that will receive the response |
|
op->_callback_response_q = callback_response_q; |
|
// user data for callback |
|
op->_callback_ptr = callback_ptr; |
|
// I am the originator of this request |
|
op->_callback_request_q = this; |
| |
op->unlock(); | op->unlock(); |
if (op->_op_dest == 0) | if (op->_op_dest == 0) |
|
|
op->_callback_parameter = parameter; | op->_callback_parameter = parameter; |
op->_callback_response_q = this; | op->_callback_response_q = this; |
| |
if (!(msg->getMask() & message_mask::ha_async)) |
if (!(msg->getMask() & MessageMask::ha_async)) |
{ | { |
AsyncLegacyOperationStart *wrapper = new AsyncLegacyOperationStart( | AsyncLegacyOperationStart *wrapper = new AsyncLegacyOperationStart( |
op, | op, |
|
|
AsyncOpNode *op = 0; | AsyncOpNode *op = 0; |
Uint32 mask = msg->getMask(); | Uint32 mask = msg->getMask(); |
| |
if (mask & message_mask::ha_async) |
if (mask & MessageMask::ha_async) |
{ | { |
op = (static_cast<AsyncMessage *>(msg))->op ; | op = (static_cast<AsyncMessage *>(msg))->op ; |
} | } |
|
|
{ | { |
op = get_op(); | op = get_op(); |
op->_request.reset(msg); | op->_request.reset(msg); |
if (mask & message_mask::ha_async) |
if (mask & MessageMask::ha_async) |
{ | { |
(static_cast<AsyncMessage *>(msg))->op = op; | (static_cast<AsyncMessage *>(msg))->op = op; |
} | } |
|
|
| |
if (reply != 0) | if (reply != 0) |
{ | { |
if (reply->getMask() & message_mask::ha_async) |
if (reply->getMask() & MessageMask::ha_async) |
{ | { |
if (reply->getMask() & message_mask::ha_reply) |
if (reply->getMask() & MessageMask::ha_reply) |
{ | { |
if (reply->result == async_results::OK || | if (reply->result == async_results::OK || |
reply->result == async_results::MODULE_ALREADY_REGISTERED) | reply->result == async_results::MODULE_ALREADY_REGISTERED) |
|
|
AsyncMessage *reply = SendWait(msg); | AsyncMessage *reply = SendWait(msg); |
if (reply) | if (reply) |
{ | { |
if (reply->getMask() & message_mask::ha_async) |
if (reply->getMask() & MessageMask::ha_async) |
{ | { |
if (reply->getMask() & message_mask::ha_reply) |
if (reply->getMask() & MessageMask::ha_reply) |
{ | { |
if (static_cast<AsyncReply *>(reply)->result == async_results::OK) |
if (static_cast<AsyncReply *>(reply)->result == |
|
async_results::OK) |
{ | { |
registered = true; | registered = true; |
} | } |
|
|
| |
Boolean MessageQueueService::deregister_service() | Boolean MessageQueueService::deregister_service() |
{ | { |
|
|
_meta_dispatcher->deregister_module(_queueId); | _meta_dispatcher->deregister_module(_queueId); |
return true; | return true; |
} | } |
|
|
AsyncMessage *reply = SendWait(req); | AsyncMessage *reply = SendWait(req); |
if (reply) | if (reply) |
{ | { |
if (reply->getMask() & message_mask::ha_async) |
if (reply->getMask() & MessageMask::ha_async) |
{ | { |
if (reply->getMask() & message_mask::ha_reply) |
if (reply->getMask() & MessageMask::ha_reply) |
{ | { |
if (reply->getType() == async_messages::FIND_SERVICE_Q_RESULT) |
if (reply->getType() == ASYNC_FIND_SERVICE_Q_RESULT) |
{ | { |
if ((static_cast<FindServiceQueueResult *>(reply))->result == async_results::OK) |
if ((static_cast<FindServiceQueueResult*>(reply))->result == |
*results = (static_cast<FindServiceQueueResult *>(reply))->qids; |
async_results::OK) |
|
*results = |
|
(static_cast<FindServiceQueueResult*>(reply))->qids; |
} | } |
} | } |
} | } |
delete reply; | delete reply; |
} | } |
delete req; | delete req; |
return ; |
|
} | } |
| |
void MessageQueueService::enumerate_service(Uint32 queue, message_module *result) |
void MessageQueueService::enumerate_service( |
|
Uint32 queue, |
|
message_module* result) |
{ | { |
if (result == 0) | if (result == 0) |
{ | { |
|
|
{ | { |
Boolean found = false; | Boolean found = false; |
| |
if (reply->getMask() & message_mask::ha_async) |
if (reply->getMask() & MessageMask::ha_async) |
{ | { |
if (reply->getMask() & message_mask::ha_reply) |
if (reply->getMask() & MessageMask::ha_reply) |
{ | { |
if (reply->getType() == async_messages::ENUMERATE_SERVICE_RESULT) |
if (reply->getType() == ASYNC_ENUMERATE_SERVICE_RESULT) |
{ | { |
if ((static_cast<EnumerateServiceResponse *>(reply))->result == async_results::OK) |
if ((static_cast<EnumerateServiceResponse*>(reply))-> |
|
result == async_results::OK) |
{ | { |
if (found == false) | if (found == false) |
{ | { |
found = true; | found = true; |
| |
result->put_name((static_cast<EnumerateServiceResponse *>(reply))->name); |
result->put_name((static_cast< |
result->put_capabilities((static_cast<EnumerateServiceResponse *>(reply))->capabilities); |
EnumerateServiceResponse*>(reply))->name); |
result->put_mask((static_cast<EnumerateServiceResponse *>(reply))->mask); |
result->put_capabilities((static_cast< |
result->put_queue((static_cast<EnumerateServiceResponse *>(reply))->qid); |
EnumerateServiceResponse*>(reply))-> |
|
capabilities); |
|
result->put_mask((static_cast< |
|
EnumerateServiceResponse*>(reply))->mask); |
|
result->put_queue((static_cast< |
|
EnumerateServiceResponse*>(reply))->qid); |
} | } |
} | } |
} | } |
|
|
delete reply; | delete reply; |
} | } |
delete req; | delete req; |
|
|
return; |
|
} | } |
| |
MessageQueueService::PollingList* MessageQueueService::_get_polling_list() | MessageQueueService::PollingList* MessageQueueService::_get_polling_list() |