version 1.4, 2002/01/21 02:01:22
|
version 1.110, 2005/06/08 04:30:58
|
|
|
//%////-*-c++-*-//////////////////////////////////////////////////////////////// |
//%2005//////////////////////////////////////////////////////////////////////// |
// | // |
// Copyright (c) 2000, 2001 The Open group, BMC Software, Tivoli Systems, IBM |
// Copyright (c) 2000, 2001, 2002 BMC Software; Hewlett-Packard Development |
|
// Company, L.P.; IBM Corp.; The Open Group; Tivoli Systems. |
|
// Copyright (c) 2003 BMC Software; Hewlett-Packard Development Company, L.P.; |
|
// IBM Corp.; EMC Corporation, The Open Group. |
|
// Copyright (c) 2004 BMC Software; Hewlett-Packard Development Company, L.P.; |
|
// IBM Corp.; EMC Corporation; VERITAS Software Corporation; The Open Group. |
|
// Copyright (c) 2005 Hewlett-Packard Development Company, L.P.; IBM Corp.; |
|
// EMC Corporation; VERITAS Software Corporation; The Open Group. |
// | // |
// Permission is hereby granted, free of charge, to any person obtaining a copy | // Permission is hereby granted, free of charge, to any person obtaining a copy |
// of this software and associated documentation files (the "Software"), to | // of this software and associated documentation files (the "Software"), to |
|
|
// Author: Mike Day (mdday@us.ibm.com) | // Author: Mike Day (mdday@us.ibm.com) |
// | // |
// Modified By: | // Modified By: |
|
// Amit K Arora, IBM (amita@in.ibm.com) for Bug#1090,#2657 |
|
// Josephine Eskaline Joyce, IBM (jojustin@in.ibm.com) for Bug#3259 |
// | // |
//%///////////////////////////////////////////////////////////////////////////// | //%///////////////////////////////////////////////////////////////////////////// |
| |
|
#include <iostream.h> |
#include "MessageQueueService.h" | #include "MessageQueueService.h" |
|
#include <Pegasus/Common/Tracer.h> |
|
#include <Pegasus/Common/MessageLoader.h> //l10n |
| |
PEGASUS_NAMESPACE_BEGIN | PEGASUS_NAMESPACE_BEGIN |
| |
|
cimom *MessageQueueService::_meta_dispatcher = 0; |
|
AtomicInt MessageQueueService::_service_count = 0; |
|
AtomicInt MessageQueueService::_xid(1); |
|
Mutex MessageQueueService::_meta_dispatcher_mutex; |
|
|
|
static struct timeval deallocateWait = {300, 0}; |
|
|
|
ThreadPool *MessageQueueService::_thread_pool = 0; |
|
|
|
DQueue<MessageQueueService> MessageQueueService::_polling_list(true); |
|
|
|
Thread* MessageQueueService::_polling_thread = 0; |
|
|
|
ThreadPool *MessageQueueService::get_thread_pool() |
|
{ |
|
return _thread_pool; |
|
} |
|
// |
|
// MAX_THREADS_PER_SVC_QUEUE_LIMIT |
|
// |
|
// 5000 is seriously too high a number for the limit but since |
|
// previously there was no limit at all this is intended to approximate |
|
// that behavior. In my testing on a unit processor system the system |
|
// behaved best with a low number 2 to 5 for the MAX_THREADS_PER_SVC_QUEUE. |
|
// When set to 1000 the system deadlocked with indications that were |
|
// not delivered and apparently left sitting within the server in a queue. |
|
// |
|
// JR Wunderlich Jun 6, 2005 |
|
// |
|
|
|
#define MAX_THREADS_PER_SVC_QUEUE_LIMIT 5000 |
|
|
|
Uint32 max_threads_per_svc_queue; |
|
|
|
PEGASUS_THREAD_RETURN PEGASUS_THREAD_CDECL |
|
MessageQueueService::kill_idle_threads(void *parm) |
|
{ |
|
|
|
static struct timeval now, last = {0,0}; |
|
gettimeofday(&now, NULL); |
|
int dead_threads = 0; |
|
|
|
if (now.tv_sec - last.tv_sec > 120) |
|
{ |
|
gettimeofday(&last, NULL); |
|
try |
|
{ |
|
dead_threads = MessageQueueService::_thread_pool->cleanupIdleThreads(); |
|
} |
|
catch(...) |
|
{ |
|
|
|
} |
|
} |
|
|
|
#ifdef PEGASUS_POINTER_64BIT |
|
return (PEGASUS_THREAD_RETURN)(Uint64)dead_threads; |
|
#elif PEGASUS_PLATFORM_AIX_RS_IBMCXX |
|
return (PEGASUS_THREAD_RETURN)(unsigned long)dead_threads; |
|
#else |
|
return (PEGASUS_THREAD_RETURN)(Uint32)dead_threads; |
|
#endif |
|
} |
| |
PEGASUS_THREAD_RETURN PEGASUS_THREAD_CDECL MessageQueueService::_req_proc(void * parm) |
PEGASUS_THREAD_RETURN PEGASUS_THREAD_CDECL MessageQueueService::polling_routine(void *parm) |
{ | { |
Thread *myself = reinterpret_cast<Thread *>(parm); | Thread *myself = reinterpret_cast<Thread *>(parm); |
MessageQueueService *service = reinterpret_cast<MessageQueueService *>(myself->get_parm()); |
DQueue<MessageQueueService> *list = reinterpret_cast<DQueue<MessageQueueService> *>(myself->get_parm()); |
|
while (_stop_polling.value() == 0) |
|
{ |
|
_polling_sem.wait(); |
| |
// pull messages off the incoming queue and dispatch them. then |
if (_stop_polling.value() != 0) |
// check pending messages that are non-blocking |
{ |
|
break; |
|
} |
| |
while ( service->_die.value() < 1 ) |
list->lock(); |
|
int list_index = 0; |
|
MessageQueueService *service = list->next(0); |
|
while(service != NULL) |
{ | { |
|
int rtn; |
|
rtn = true; |
|
if (service->_incoming.count() > 0 |
|
&& service->_die.value() == 0 |
|
&& service->_threads <= max_threads_per_svc_queue) |
|
rtn = _thread_pool->allocate_and_awaken(service, _req_proc, |
|
&_polling_sem); |
| |
AsyncOpNode *operation = service->_incoming.remove_first_wait(); |
// if no more threads available, break from processing loop |
service->_handle_incoming_operation(operation); |
if (rtn == false) |
|
{ |
|
service = NULL; |
|
} |
|
else |
|
{ |
|
service = list->next(service); |
} | } |
|
} |
|
list->unlock(); |
|
|
|
if (_check_idle_flag.value() != 0) |
|
{ |
|
_check_idle_flag = 0; |
|
// try to do idle thread clean up processing when system is not busy |
|
// if system is busy there may not be a thread available to allocate |
|
// so nothing will be done and that is OK. |
| |
|
_thread_pool->allocate_and_awaken(service, kill_idle_threads, &_polling_sem); |
|
|
|
} |
|
} |
myself->exit_self( (PEGASUS_THREAD_RETURN) 1 ); | myself->exit_self( (PEGASUS_THREAD_RETURN) 1 ); |
return(0); | return(0); |
} | } |
| |
PEGASUS_THREAD_RETURN PEGASUS_THREAD_CDECL MessageQueueService::_rpl_proc(void * parm) |
|
|
Semaphore MessageQueueService::_polling_sem(0); |
|
AtomicInt MessageQueueService::_stop_polling(0); |
|
AtomicInt MessageQueueService::_check_idle_flag(0); |
|
|
|
|
|
MessageQueueService::MessageQueueService( |
|
const char *name, |
|
Uint32 queueID, |
|
Uint32 capabilities, |
|
Uint32 mask) |
|
: Base(name, true, queueID), |
|
_mask(mask), |
|
_die(0), |
|
_threads(0), |
|
_incoming(true, 0), |
|
_incoming_queue_shutdown(0) |
{ | { |
Thread *myself = reinterpret_cast<Thread *>(parm); |
|
MessageQueueService *service = reinterpret_cast<MessageQueueService *>(myself->get_parm()); |
|
| |
// pull messages off the _pending queue and complete them. |
_capabilities = (capabilities | module_capabilities::async); |
// this loop is not optimized because every time we complete an operation |
|
// we start again to traverse the loop at the beginning. |
_default_op_timeout.tv_sec = 30; |
Uint32 loop = 1; |
_default_op_timeout.tv_usec = 100; |
Boolean unlinked = false; |
|
|
max_threads_per_svc_queue = MAX_THREADS_PER_SVC_QUEUE; |
| |
|
// if requested threads gt MAX_THREADS_PER_SVC_QUEUE_LIMIT |
|
// then set to MAX_THREADS_PER_SVC_QUEUE_LIMIT |
| |
while ( service->_die.value() < 1 ) |
if (max_threads_per_svc_queue > MAX_THREADS_PER_SVC_QUEUE_LIMIT) |
{ | { |
service->_pending.wait_for_node(); |
max_threads_per_svc_queue = MAX_THREADS_PER_SVC_QUEUE_LIMIT; |
AsyncOpNode *operation = service->_pending.next(0); |
} |
// operation = service->_pending.remove_no_lock(operation); |
|
while( operation != 0 ) |
// if requested threads eq 0 (unlimited) |
|
// then set to MAX_THREADS_PER_SVC_QUEUE_LIMIT |
|
|
|
if (max_threads_per_svc_queue == 0) |
{ | { |
if ( operation->_user_data != loop ) |
max_threads_per_svc_queue = MAX_THREADS_PER_SVC_QUEUE_LIMIT; |
|
} |
|
|
|
// cout << "MAX_THREADS_PER_SVC_QUEUE = " << MAX_THREADS_PER_SVC_QUEUE << endl; |
|
// cout << "max_threads_per_svc_queue set to = " << max_threads_per_svc_queue << endl; |
|
|
|
|
|
AutoMutex autoMut(_meta_dispatcher_mutex); |
|
|
|
if (_meta_dispatcher == 0) |
{ | { |
Uint32 state = operation->read_state(); |
_stop_polling = 0; |
|
PEGASUS_ASSERT(_service_count.value() == 0); |
|
_meta_dispatcher = new cimom(); |
|
if (_meta_dispatcher == NULL) |
|
{ |
|
throw NullPointer(); |
|
} |
|
// _thread_pool = new ThreadPool(initial_cnt, "MessageQueueService", |
|
// minimum_cnt, maximum_cnt, deallocateWait); |
|
// |
|
_thread_pool = |
|
new ThreadPool(0, "MessageQueueService", 0, 0, deallocateWait); |
|
} |
|
_service_count++; |
| |
if ( state & ASYNC_OPSTATE_COMPLETE ) |
if (false == register_service(name, _capabilities, _mask)) |
{ | { |
service->_pending.remove_no_lock(operation) ; |
//l10n |
operation->_client_sem.signal(); |
//throw BindFailedException("MessageQueueService Base Unable to register with Meta Dispatcher"); |
operation = service->_pending.next(0); |
MessageLoaderParms parms("Common.MessageQueueService.UNABLE_TO_REGISTER", |
unlinked = true; |
"MessageQueueService Base Unable to register with Meta Dispatcher"); |
continue; |
|
|
throw BindFailedException(parms); |
} | } |
operation->_user_data = loop; |
|
|
_polling_list.insert_last(this); |
|
|
} | } |
operation = service->_pending.next(operation); |
|
|
|
|
MessageQueueService::~MessageQueueService() |
|
{ |
|
_die = 1; |
|
|
|
if (_incoming_queue_shutdown.value() == 0) |
|
{ |
|
_shutdown_incoming_queue(); |
|
|
} | } |
service->_pending.unlock(); |
|
loop++; |
while (_threads.value() > 0) |
if ( unlinked == true) |
|
{ | { |
pegasus_yield(); | pegasus_yield(); |
unlinked = false; |
|
} |
|
} | } |
|
_polling_list.remove(this); |
|
{ |
|
AutoMutex autoMut(_meta_dispatcher_mutex); |
|
_service_count--; |
|
if (_service_count.value() == 0) |
|
{ |
| |
myself->exit_self( (PEGASUS_THREAD_RETURN) 1 ); |
_stop_polling++; |
return(0); |
_polling_sem.signal(); |
|
if (_polling_thread) { |
|
_polling_thread->join(); |
|
delete _polling_thread; |
|
_polling_thread = 0; |
} | } |
|
_meta_dispatcher->_shutdown_routed_queue(); |
|
delete _meta_dispatcher; |
|
_meta_dispatcher = 0; |
| |
|
delete _thread_pool; |
void MessageQueueService::_handle_incoming_operation(AsyncOpNode *operation) |
_thread_pool = 0; |
|
} |
|
} // mutex unlocks here |
|
// Clean up in case there are extra stuff on the queue. |
|
while (_incoming.count()) |
{ | { |
if ( operation != 0 ) |
try { |
|
delete _incoming.remove_first(); |
|
} catch (const ListClosed &e) |
{ | { |
Message *rq = operation->get_request(); |
// If the list is closed, there is nothing we can do. |
PEGASUS_ASSERT(rq != 0 ); |
break; |
PEGASUS_ASSERT(rq->getMask() & message_mask::ha_async ); |
} |
PEGASUS_ASSERT(rq->getMask() & message_mask::ha_request); |
} |
_handle_async_request(static_cast<AsyncRequest *>(rq)); |
|
} | } |
| |
|
void MessageQueueService::_shutdown_incoming_queue() |
|
{ |
|
if (_incoming_queue_shutdown.value() > 0) |
return; | return; |
| |
|
AsyncIoctl *msg = new AsyncIoctl( |
|
get_next_xid(), |
|
0, |
|
_queueId, |
|
_queueId, |
|
true, |
|
AsyncIoctl::IO_CLOSE, |
|
0, |
|
0); |
|
|
|
msg->op = get_op(); |
|
msg->op->_flags |= ASYNC_OPFLAGS_FIRE_AND_FORGET; |
|
msg->op->_flags &= ~(ASYNC_OPFLAGS_CALLBACK | ASYNC_OPFLAGS_SAFE_CALLBACK |
|
| ASYNC_OPFLAGS_SIMPLE_STATUS); |
|
msg->op->_state &= ~ASYNC_OPSTATE_COMPLETE; |
|
|
|
msg->op->_op_dest = this; |
|
msg->op->_request.insert_first(msg); |
|
try { |
|
_incoming.insert_last_wait(msg->op); |
|
_polling_sem.signal(); |
|
} catch (const ListClosed &) |
|
{ |
|
// This means the queue has already been shut-down (happens when there |
|
// are two AsyncIoctrl::IO_CLOSE messages generated and one got first |
|
// processed. |
|
delete msg; |
|
} |
} | } |
| |
| |
Boolean MessageQueueService::messageOK(const Message *msg) |
|
{ |
void MessageQueueService::enqueue(Message *msg) |
if ( msg != 0 ) |
|
{ | { |
Uint32 mask = msg->getMask(); |
PEG_METHOD_ENTER(TRC_MESSAGEQUEUESERVICE, "MessageQueueService::enqueue()"); |
if ( mask & message_mask::ha_async) |
|
if ( mask & message_mask::ha_request) |
Base::enqueue(msg); |
return true; |
|
} |
PEG_METHOD_EXIT(); |
return false; |
|
} | } |
| |
| |
MessageQueueService::MessageQueueService(const char *name, |
PEGASUS_THREAD_RETURN PEGASUS_THREAD_CDECL MessageQueueService::_req_proc( |
Uint32 queueID, |
void * parm) |
Uint32 capabilities, |
{ |
Uint32 mask) |
MessageQueueService* service = |
: Base(name, false, queueID), |
reinterpret_cast<MessageQueueService*>(parm); |
_capabilities(capabilities), |
PEGASUS_ASSERT(service != 0); |
_mask(mask), |
try |
_die(0), |
|
_pending(true, 1000), |
|
_incoming(true, 1000), |
|
_req_thread(_req_proc, this, false), |
|
_rpl_thread(_rpl_proc, this, false) |
|
{ | { |
_default_op_timeout.tv_sec = 30; |
|
_default_op_timeout.tv_usec = 100; |
|
_meta_dispatcher = static_cast<cimom *>(Base::lookup(CIMOM_Q_ID)); |
|
if(_meta_dispatcher == 0 ) |
|
throw NullPointer(); |
|
_req_thread.run(); |
|
_rpl_thread.run(); |
|
| |
|
if (service->_die.value() != 0) |
|
{ |
|
return (0); |
} | } |
|
service->_threads++; |
|
// pull messages off the incoming queue and dispatch them. then |
|
// check pending messages that are non-blocking |
|
AsyncOpNode *operation = 0; |
| |
|
// many operations may have been queued. |
|
do |
|
{ |
|
try |
|
{ |
|
operation = service->_incoming.remove_first(); |
|
} |
|
catch (ListClosed &) |
|
{ |
|
// ATTN: This appears to be a common loop exit path. |
|
//PEG_TRACE_STRING(TRC_DISCARDED_DATA, Tracer::LEVEL2, |
|
// "Caught ListClosed exception. Exiting _req_proc."); |
|
break; |
|
} |
| |
MessageQueueService::~MessageQueueService(void) |
if (operation) |
{ | { |
_die = 1; |
operation->_service_ptr = service; |
_pending.shutdown_queue(); |
service->_handle_incoming_operation(operation); |
_incoming.shutdown_queue(); |
} |
_req_thread.join(); |
} while (operation); |
_rpl_thread.join(); |
} |
|
catch (const Exception& e) |
|
{ |
|
PEG_TRACE_STRING(TRC_DISCARDED_DATA, Tracer::LEVEL2, |
|
String("Caught exception: \"") + e.getMessage() + |
|
"\". Exiting _req_proc."); |
|
} |
|
catch (...) |
|
{ |
|
PEG_TRACE_STRING(TRC_DISCARDED_DATA, Tracer::LEVEL2, |
|
"Caught unrecognized exception. Exiting _req_proc."); |
|
} |
|
service->_threads--; |
|
return(0); |
} | } |
| |
AtomicInt MessageQueueService::_xid(1); |
|
|
void MessageQueueService::_sendwait_callback( |
|
AsyncOpNode *op, |
|
MessageQueue *q, |
|
void *parm) |
|
{ |
|
op->_client_sem.signal(); |
|
} |
| |
| |
Boolean MessageQueueService::accept_async(AsyncOpNode *op) throw(IPCException) |
// callback function is responsible for cleaning up all resources |
|
// including op, op->_callback_node, and op->_callback_ptr |
|
void MessageQueueService::_handle_async_callback(AsyncOpNode *op) |
|
{ |
|
if (op->_flags & ASYNC_OPFLAGS_SAFE_CALLBACK) |
{ | { |
// incoming async request message |
|
| |
PEGASUS_ASSERT(op != 0 ); |
Message *msg = op->get_request(); |
|
if (msg && (msg->getMask() & message_mask::ha_async)) |
|
{ |
|
if (msg->getType() == async_messages::ASYNC_LEGACY_OP_START) |
|
{ |
|
AsyncLegacyOperationStart *wrapper = |
|
static_cast<AsyncLegacyOperationStart *>(msg); |
|
msg = wrapper->get_action(); |
|
delete wrapper; |
|
} |
|
else if (msg->getType() == async_messages::ASYNC_MODULE_OP_START) |
|
{ |
|
AsyncModuleOperationStart *wrapper = |
|
static_cast<AsyncModuleOperationStart *>(msg); |
|
msg = wrapper->get_action(); |
|
delete wrapper; |
|
} |
|
else if (msg->getType() == async_messages::ASYNC_OP_START) |
|
{ |
|
AsyncOperationStart *wrapper = |
|
static_cast<AsyncOperationStart *>(msg); |
|
msg = wrapper->get_action(); |
|
delete wrapper; |
|
} |
|
delete msg; |
|
} |
| |
Message *rq = op->get_request(); |
msg = op->get_response(); |
if ( true == messageOK(rq) ) |
if (msg && (msg->getMask() & message_mask::ha_async)) |
{ | { |
try |
if (msg->getType() == async_messages::ASYNC_LEGACY_OP_RESULT) |
{ | { |
_incoming.insert_first_wait(op); |
AsyncLegacyOperationResult *wrapper = |
|
static_cast<AsyncLegacyOperationResult *>(msg); |
|
msg = wrapper->get_result(); |
|
delete wrapper; |
} | } |
catch ( IPCException& e) |
else if (msg->getType() == async_messages::ASYNC_MODULE_OP_RESULT) |
{ | { |
return false ; |
AsyncModuleOperationResult *wrapper = |
|
static_cast<AsyncModuleOperationResult *>(msg); |
|
msg = wrapper->get_result(); |
|
delete wrapper; |
} | } |
return true; |
|
} | } |
return false; |
void (*callback)(Message *, void *, void *) = op->__async_callback; |
|
void *handle = op->_callback_handle; |
|
void *parm = op->_callback_parameter; |
|
op->release(); |
|
return_op(op); |
|
callback(msg, handle, parm); |
|
} |
|
else if (op->_flags & ASYNC_OPFLAGS_CALLBACK) |
|
{ |
|
// note that _callback_node may be different from op |
|
// op->_callback_response_q is a "this" pointer we can use for |
|
// static callback methods |
|
op->_async_callback(op->_callback_node, op->_callback_response_q, op->_callback_ptr); |
|
} |
} | } |
| |
| |
void MessageQueueService::_completeAsyncResponse(AsyncRequest *request, |
void MessageQueueService::_handle_incoming_operation(AsyncOpNode *operation) |
AsyncReply *reply, |
{ |
Uint32 state, |
if (operation != 0) |
Uint32 flag) |
|
{ | { |
PEGASUS_ASSERT(request != 0 && reply != 0 ); |
|
| |
AsyncOpNode *op = request->op; |
// ATTN: optimization |
op->lock(); |
// << Tue Feb 19 14:10:38 2002 mdd >> |
// if (false == op->_response.exists(reply)) |
operation->lock(); |
// op->_response.insert_last(reply); |
|
op->_response = reply; |
Message *rq = operation->_request.next(0); |
|
|
op->_state |= state ; |
// optimization <<< Thu Mar 7 21:04:05 2002 mdd >>> |
op->_flags |= flag; |
// move this to the bottom of the loop when the majority of |
gettimeofday(&(op->_updated), NULL); |
// messages become async messages. |
op->unlock(); |
|
|
// divert legacy messages to handleEnqueue |
|
if ((rq != 0) && (!(rq->getMask() & message_mask::ha_async))) |
|
{ |
|
rq = operation->_request.remove_first() ; |
|
operation->unlock(); |
|
// delete the op node |
|
operation->release(); |
|
return_op(operation); |
|
|
|
handleEnqueue(rq); |
|
return; |
} | } |
| |
void MessageQueueService::handleEnqueue(void) |
if ((operation->_flags & ASYNC_OPFLAGS_CALLBACK || |
|
operation->_flags & ASYNC_OPFLAGS_SAFE_CALLBACK) && |
|
(operation->_state & ASYNC_OPSTATE_COMPLETE)) |
{ | { |
Message *msg = dequeue(); |
operation->unlock(); |
if( msg ) |
_handle_async_callback(operation); |
|
} |
|
else |
{ | { |
delete msg; |
PEGASUS_ASSERT(rq != 0); |
|
operation->unlock(); |
|
_handle_async_request(static_cast<AsyncRequest *>(rq)); |
} | } |
} | } |
|
return; |
|
} |
| |
void MessageQueueService::_handle_async_request(AsyncRequest *req) | void MessageQueueService::_handle_async_request(AsyncRequest *req) |
{ | { |
|
|
} | } |
} | } |
| |
void MessageQueueService::_handle_async_reply(AsyncReply *rep) |
|
|
Boolean MessageQueueService::_enqueueResponse( |
|
Message* request, |
|
Message* response) |
{ | { |
| |
if (rep->op != 0 ) |
STAT_COPYDISPATCHER |
rep->op->processing(); |
|
| |
Uint32 type = rep->getType(); |
PEG_METHOD_ENTER(TRC_MESSAGEQUEUESERVICE, |
|
"MessageQueueService::_enqueueResponse"); |
| |
if ( type == async_messages::ASYNC_OP_RESULT ) |
if (request->getMask() & message_mask::ha_async) |
handle_AsyncOperationResult(static_cast<AsyncOperationResult *>(rep)); |
|
else |
|
{ | { |
// we don't handle this reply |
if (response->getMask() & message_mask::ha_async) |
; |
{ |
|
_completeAsyncResponse(static_cast<AsyncRequest *>(request), |
|
static_cast<AsyncReply *>(response), |
|
ASYNC_OPSTATE_COMPLETE, 0); |
|
PEG_METHOD_EXIT(); |
|
return true; |
|
} |
|
} |
|
|
|
if (request->_async != 0) |
|
{ |
|
Uint32 mask = request->_async->getMask(); |
|
PEGASUS_ASSERT(mask & (message_mask::ha_async | message_mask::ha_request)); |
|
|
|
AsyncRequest *async = static_cast<AsyncRequest *>(request->_async); |
|
AsyncOpNode *op = async->op; |
|
request->_async = 0; |
|
// the legacy request is going to be deleted by its handler |
|
// remove it from the op node |
|
|
|
static_cast<AsyncLegacyOperationStart *>(async)->get_action(); |
|
|
|
AsyncLegacyOperationResult *async_result = |
|
new AsyncLegacyOperationResult( |
|
async->getKey(), |
|
async->getRouting(), |
|
op, |
|
response); |
|
_completeAsyncResponse( |
|
async, |
|
async_result, |
|
ASYNC_OPSTATE_COMPLETE, |
|
0); |
|
PEG_METHOD_EXIT(); |
|
return true; |
} | } |
| |
if( rep->op != 0 ) |
// ensure that the destination queue is in response->dest |
rep->op->release(); |
PEG_METHOD_EXIT(); |
|
return SendForget(response); |
|
|
} | } |
| |
void MessageQueueService::_make_response(AsyncRequest *req, Uint32 code) |
void MessageQueueService::_make_response(Message *req, Uint32 code) |
{ | { |
AsyncReply *reply = |
cimom::_make_response(req, code); |
new AsyncReply(async_messages::REPLY, |
} |
req->getKey(), |
|
req->getRouting(), |
|
0, |
void MessageQueueService::_completeAsyncResponse( |
req->op, |
AsyncRequest *request, |
code, |
AsyncReply *reply, |
req->resp, |
Uint32 state, |
|
Uint32 flag) |
|
{ |
|
PEG_METHOD_ENTER(TRC_MESSAGEQUEUESERVICE, |
|
"MessageQueueService::_completeAsyncResponse"); |
|
|
|
cimom::_completeAsyncResponse(request, reply, state, flag); |
|
|
|
PEG_METHOD_EXIT(); |
|
} |
|
|
|
|
|
void MessageQueueService::_complete_op_node( |
|
AsyncOpNode *op, |
|
Uint32 state, |
|
Uint32 flag, |
|
Uint32 code) |
|
{ |
|
cimom::_complete_op_node(op, state, flag, code); |
|
} |
|
|
|
|
|
Boolean MessageQueueService::accept_async(AsyncOpNode *op) |
|
{ |
|
if (_incoming_queue_shutdown.value() > 0) |
|
return false; |
|
if (_polling_thread == NULL) |
|
{ |
|
_polling_thread = new Thread( |
|
polling_routine, |
|
reinterpret_cast<void *>(&_polling_list), |
false); | false); |
_completeAsyncResponse(req, reply, ASYNC_OPSTATE_COMPLETE, 0 ); |
while (!_polling_thread->run()) |
|
{ |
|
pegasus_yield(); |
} | } |
|
} |
|
// ATTN optimization remove the message checking altogether in the base |
|
// << Mon Feb 18 14:02:20 2002 mdd >> |
|
op->lock(); |
|
Message *rq = op->_request.next(0); |
|
Message *rp = op->_response.next(0); |
|
op->unlock(); |
| |
|
if ((rq != 0 && (true == messageOK(rq))) || |
|
(rp != 0 && (true == messageOK(rp))) && _die.value() == 0) |
|
{ |
|
_incoming.insert_last_wait(op); |
|
_polling_sem.signal(); |
|
return true; |
|
} |
|
return false; |
|
} |
|
|
|
Boolean MessageQueueService::messageOK(const Message *msg) |
|
{ |
|
if (_incoming_queue_shutdown.value() > 0) |
|
return false; |
|
return true; |
|
} |
| |
void MessageQueueService::handle_heartbeat_request(AsyncRequest *req) | void MessageQueueService::handle_heartbeat_request(AsyncRequest *req) |
{ | { |
// default action is to echo a heartbeat response | // default action is to echo a heartbeat response |
| |
AsyncReply *reply = |
AsyncReply *reply = new AsyncReply( |
new AsyncReply(async_messages::HEARTBEAT, |
async_messages::HEARTBEAT, |
req->getKey(), | req->getKey(), |
req->getRouting(), | req->getRouting(), |
0, | 0, |
|
|
| |
void MessageQueueService::handle_heartbeat_reply(AsyncReply *rep) | void MessageQueueService::handle_heartbeat_reply(AsyncReply *rep) |
{ | { |
; |
|
} | } |
| |
void MessageQueueService::handle_AsyncIoctl(AsyncIoctl *req) | void MessageQueueService::handle_AsyncIoctl(AsyncIoctl *req) |
{ | { |
|
switch (req->ctl) |
|
{ |
|
case AsyncIoctl::IO_CLOSE: |
|
{ |
|
MessageQueueService *service = static_cast<MessageQueueService *>(req->op->_service_ptr); |
|
|
|
#ifdef MESSAGEQUEUESERVICE_DEBUG |
|
PEGASUS_STD(cout) << service->getQueueName() << " Received AsyncIoctl::IO_CLOSE " << PEGASUS_STD(endl); |
|
#endif |
|
|
|
// respond to this message. this is fire and forget, so we don't need to delete anything. |
|
// this takes care of two problems that were being found |
|
// << Thu Oct 9 10:52:48 2003 mdd >> |
_make_response(req, async_results::OK); | _make_response(req, async_results::OK); |
|
// ensure we do not accept any further messages |
|
|
|
// ensure we don't recurse on IO_CLOSE |
|
if (_incoming_queue_shutdown.value() > 0) |
|
break; |
|
|
|
// set the closing flag |
|
service->_incoming_queue_shutdown = 1; |
|
// empty out the queue |
|
while (1) |
|
{ |
|
AsyncOpNode *operation; |
|
try |
|
{ |
|
operation = service->_incoming.remove_first(); |
} | } |
void MessageQueueService::handle_CimServiceStart(CimServiceStart *req) |
catch(IPCException &) |
|
{ |
|
break; |
|
} |
|
if (operation) |
{ | { |
|
operation->_service_ptr = service; |
|
service->_handle_incoming_operation(operation); |
|
} |
|
else |
|
break; |
|
} // message processing loop |
|
|
|
// shutdown the AsyncDQueue |
|
service->_incoming.shutdown_queue(); |
|
return; |
|
} |
|
|
|
default: |
_make_response(req, async_results::CIM_NAK); | _make_response(req, async_results::CIM_NAK); |
} | } |
|
} |
|
|
|
void MessageQueueService::handle_CimServiceStart(CimServiceStart *req) |
|
{ |
|
|
|
#ifdef MESSAGEQUEUESERVICE_DEBUG |
|
PEGASUS_STD(cout) << getQueueName() << "received START" << PEGASUS_STD(endl); |
|
#endif |
|
|
|
// clear the stoped bit and update |
|
_capabilities &= (~(module_capabilities::stopped)); |
|
_make_response(req, async_results::OK); |
|
// now tell the meta dispatcher we are stopped |
|
update_service(_capabilities, _mask); |
|
|
|
} |
void MessageQueueService::handle_CimServiceStop(CimServiceStop *req) | void MessageQueueService::handle_CimServiceStop(CimServiceStop *req) |
{ | { |
_make_response(req, async_results::CIM_NAK); |
#ifdef MESSAGEQUEUESERVICE_DEBUG |
|
PEGASUS_STD(cout) << getQueueName() << "received STOP" << PEGASUS_STD(endl); |
|
#endif |
|
// set the stopeed bit and update |
|
_capabilities |= module_capabilities::stopped; |
|
_make_response(req, async_results::CIM_STOPPED); |
|
// now tell the meta dispatcher we are stopped |
|
update_service(_capabilities, _mask); |
} | } |
|
|
void MessageQueueService::handle_CimServicePause(CimServicePause *req) | void MessageQueueService::handle_CimServicePause(CimServicePause *req) |
{ | { |
_make_response(req, async_results::CIM_NAK); |
// set the paused bit and update |
|
_capabilities |= module_capabilities::paused; |
|
update_service(_capabilities, _mask); |
|
_make_response(req, async_results::CIM_PAUSED); |
|
// now tell the meta dispatcher we are stopped |
} | } |
|
|
void MessageQueueService::handle_CimServiceResume(CimServiceResume *req) | void MessageQueueService::handle_CimServiceResume(CimServiceResume *req) |
{ | { |
_make_response(req, async_results::CIM_NAK); |
// clear the paused bit and update |
|
_capabilities &= (~(module_capabilities::paused)); |
|
update_service(_capabilities, _mask); |
|
_make_response(req, async_results::OK); |
|
// now tell the meta dispatcher we are stopped |
} | } |
| |
void MessageQueueService::handle_AsyncOperationStart(AsyncOperationStart *req) | void MessageQueueService::handle_AsyncOperationStart(AsyncOperationStart *req) |
{ | { |
_make_response(req, async_results::CIM_NAK); | _make_response(req, async_results::CIM_NAK); |
|
|
} | } |
| |
void MessageQueueService::handle_AsyncOperationResult(AsyncOperationResult *req) | void MessageQueueService::handle_AsyncOperationResult(AsyncOperationResult *req) |
|
|
; | ; |
} | } |
| |
AsyncOpNode *MessageQueueService::get_op(void) |
|
|
void MessageQueueService::handle_AsyncLegacyOperationStart(AsyncLegacyOperationStart *req) |
|
{ |
|
// remove the legacy message from the request and enqueue it to its destination |
|
Uint32 result = async_results::CIM_NAK; |
|
|
|
Message *legacy = req->_act; |
|
if (legacy != 0) |
|
{ |
|
MessageQueue* queue = MessageQueue::lookup(req->_legacy_destination); |
|
if (queue != 0) |
|
{ |
|
if (queue->isAsync() == true) |
|
{ |
|
(static_cast<MessageQueueService *>(queue))->handleEnqueue(legacy); |
|
} |
|
else |
|
{ |
|
// Enqueue the response: |
|
queue->enqueue(req->get_action()); |
|
} |
|
|
|
result = async_results::OK; |
|
} |
|
} |
|
_make_response(req, result); |
|
} |
|
|
|
void MessageQueueService::handle_AsyncLegacyOperationResult(AsyncLegacyOperationResult *rep) |
|
{ |
|
; |
|
} |
|
|
|
AsyncOpNode *MessageQueueService::get_op() |
{ | { |
AsyncOpNode *op = new AsyncOpNode(); | AsyncOpNode *op = new AsyncOpNode(); |
| |
op->write_state(ASYNC_OPSTATE_UNKNOWN); |
op->_state = ASYNC_OPSTATE_UNKNOWN; |
op->write_flags(ASYNC_OPFLAGS_SINGLE | ASYNC_OPFLAGS_NORMAL ); |
op->_flags = ASYNC_OPFLAGS_SINGLE | ASYNC_OPFLAGS_NORMAL; |
| |
return op; | return op; |
} | } |
|
|
} | } |
| |
| |
|
Boolean MessageQueueService::ForwardOp( |
|
AsyncOpNode *op, |
|
Uint32 destination) |
|
{ |
|
PEGASUS_ASSERT(op != 0); |
|
op->lock(); |
|
op->_op_dest = MessageQueue::lookup(destination); |
|
op->_flags |= (ASYNC_OPFLAGS_FIRE_AND_FORGET | ASYNC_OPFLAGS_FORWARD); |
|
op->_flags &= ~(ASYNC_OPFLAGS_CALLBACK); |
|
op->unlock(); |
|
if (op->_op_dest == 0) |
|
return false; |
| |
AsyncReply *MessageQueueService::SendWait(AsyncRequest *request) |
return _meta_dispatcher->route_async(op); |
|
} |
|
|
|
|
|
Boolean MessageQueueService::SendAsync( |
|
AsyncOpNode *op, |
|
Uint32 destination, |
|
void (*callback)(AsyncOpNode *, MessageQueue *, void *), |
|
MessageQueue *callback_response_q, |
|
void *callback_ptr) |
{ | { |
if ( request == 0 ) |
PEGASUS_ASSERT(op != 0 && callback != 0); |
return 0 ; |
|
PEGASUS_ASSERT( request->op != 0 ); |
|
| |
request->block = true; |
// get the queue handle for the destination |
request->op->_state &= ~ASYNC_OPSTATE_COMPLETE; |
|
request->op->put_response(0); |
|
| |
// first link it on our pending list |
op->lock(); |
_pending.insert_last_wait(request->op); |
op->_op_dest = MessageQueue::lookup(destination); // destination of this message |
|
op->_flags |= ASYNC_OPFLAGS_CALLBACK; |
|
op->_flags &= ~(ASYNC_OPFLAGS_FIRE_AND_FORGET); |
|
// initialize the callback data |
|
op->_async_callback = callback; // callback function to be executed by recpt. of response |
|
op->_callback_node = op; // the op node |
|
op->_callback_response_q = callback_response_q; // the queue that will receive the response |
|
op->_callback_ptr = callback_ptr; // user data for callback |
|
op->_callback_request_q = this; // I am the originator of this request |
| |
// now see if the meta dispatcher will take it |
op->unlock(); |
|
if (op->_op_dest == 0) |
|
return false; |
|
|
|
return _meta_dispatcher->route_async(op); |
|
} |
| |
if (true == _meta_dispatcher->route_async(request->op)) |
|
|
Boolean MessageQueueService::SendAsync( |
|
Message *msg, |
|
Uint32 destination, |
|
void (*callback)(Message *response, void *handle, void *parameter), |
|
void *handle, |
|
void *parameter) |
{ | { |
request->op->_client_sem.wait(); |
if (msg == NULL) |
|
return false; |
|
if (callback == NULL) |
|
return SendForget(msg); |
|
AsyncOpNode *op = get_op(); |
|
msg->dest = destination; |
|
if (NULL == (op->_op_dest = MessageQueue::lookup(msg->dest))) |
|
{ |
|
op->release(); |
|
return_op(op); |
|
return false; |
|
} |
|
op->_flags |= ASYNC_OPFLAGS_SAFE_CALLBACK; |
|
op->_flags &= ~(ASYNC_OPFLAGS_FIRE_AND_FORGET); |
|
op->_state &= ~ASYNC_OPSTATE_COMPLETE; |
|
op->__async_callback = callback; |
|
op->_callback_node = op; |
|
op->_callback_handle = handle; |
|
op->_callback_parameter = parameter; |
|
op->_callback_response_q = this; |
|
|
|
if (!(msg->getMask() & message_mask::ha_async)) |
|
{ |
|
AsyncLegacyOperationStart *wrapper = new AsyncLegacyOperationStart( |
|
get_next_xid(), |
|
op, |
|
destination, |
|
msg, |
|
destination); |
|
} |
|
else |
|
{ |
|
op->_request.insert_first(msg); |
|
(static_cast<AsyncMessage *>(msg))->op = op; |
|
} |
|
return _meta_dispatcher->route_async(op); |
|
} |
|
|
|
|
|
Boolean MessageQueueService::SendForget(Message *msg) |
|
{ |
|
AsyncOpNode *op = 0; |
|
Uint32 mask = msg->getMask(); |
|
|
|
if (mask & message_mask::ha_async) |
|
{ |
|
op = (static_cast<AsyncMessage *>(msg))->op ; |
} | } |
| |
return static_cast<AsyncReply *>(request->op->get_response()); |
if (op == 0) |
|
{ |
|
op = get_op(); |
|
op->_request.insert_first(msg); |
|
if (mask & message_mask::ha_async) |
|
{ |
|
(static_cast<AsyncMessage *>(msg))->op = op; |
|
} |
|
} |
|
op->_op_dest = MessageQueue::lookup(msg->dest); |
|
op->_flags |= ASYNC_OPFLAGS_FIRE_AND_FORGET; |
|
op->_flags &= ~(ASYNC_OPFLAGS_CALLBACK | ASYNC_OPFLAGS_SAFE_CALLBACK |
|
| ASYNC_OPFLAGS_SIMPLE_STATUS); |
|
op->_state &= ~ASYNC_OPSTATE_COMPLETE; |
|
if (op->_op_dest == 0) |
|
{ |
|
op->release(); |
|
return_op(op); |
|
return false; |
|
} |
| |
|
// now see if the meta dispatcher will take it |
|
return _meta_dispatcher->route_async(op); |
} | } |
| |
| |
Boolean MessageQueueService::register_service(String name, |
AsyncReply *MessageQueueService::SendWait(AsyncRequest *request) |
Uint32 capabilities, |
{ |
Uint32 mask) |
if (request == 0) |
|
return 0 ; |
| |
|
Boolean destroy_op = false; |
|
|
|
if (request->op == 0) |
{ | { |
AsyncOpNode *op = _meta_dispatcher->get_cached_op(); |
request->op = get_op(); |
|
request->op->_request.insert_first(request); |
|
destroy_op = true; |
|
} |
|
|
|
request->block = false; |
|
request->op->_flags |= ASYNC_OPFLAGS_PSEUDO_CALLBACK; |
|
SendAsync( |
|
request->op, |
|
request->dest, |
|
_sendwait_callback, |
|
this, |
|
(void *)0); |
| |
op->_state |= ASYNC_OPSTATE_UNKNOWN; |
request->op->_client_sem.wait(); |
op->_flags |= ASYNC_OPFLAGS_SINGLE | ASYNC_OPFLAGS_NORMAL; |
|
| |
RegisterCimService *msg = new RegisterCimService(get_next_xid(), |
request->op->lock(); |
op, |
AsyncReply * rpl = static_cast<AsyncReply *>(request->op->_response.remove_first()); |
|
rpl->op = 0; |
|
request->op->unlock(); |
|
|
|
if (destroy_op == true) |
|
{ |
|
request->op->lock(); |
|
request->op->_request.remove(request); |
|
request->op->_state |= ASYNC_OPSTATE_RELEASED; |
|
request->op->unlock(); |
|
return_op(request->op); |
|
request->op = 0; |
|
} |
|
return rpl; |
|
} |
|
|
|
|
|
Boolean MessageQueueService::register_service( |
|
String name, |
|
Uint32 capabilities, |
|
Uint32 mask) |
|
{ |
|
RegisterCimService *msg = new RegisterCimService( |
|
get_next_xid(), |
|
0, |
true, | true, |
name, | name, |
capabilities, | capabilities, |
mask, | mask, |
_queueId); | _queueId); |
|
msg->dest = CIMOM_Q_ID; |
|
|
Boolean registered = false; | Boolean registered = false; |
AsyncMessage *reply = SendWait( msg ); |
AsyncReply *reply = static_cast<AsyncReply *>(SendWait(msg)); |
| |
if ( reply != 0 ) | if ( reply != 0 ) |
{ | { |
|
|
{ | { |
if(reply->getMask() & message_mask::ha_reply) | if(reply->getMask() & message_mask::ha_reply) |
{ | { |
if((static_cast<AsyncReply *>(reply))->result == async_results::OK) |
if (reply->result == async_results::OK || |
|
reply->result == async_results::MODULE_ALREADY_REGISTERED) |
|
{ |
registered = true; | registered = true; |
|
} |
} | } |
} | } |
| |
delete reply; | delete reply; |
} | } |
|
delete msg; |
return registered; | return registered; |
} | } |
| |
Boolean MessageQueueService::update_service(Uint32 capabilities, Uint32 mask) | Boolean MessageQueueService::update_service(Uint32 capabilities, Uint32 mask) |
{ | { |
|
UpdateCimService *msg = new UpdateCimService( |
AsyncOpNode *op = _meta_dispatcher->get_cached_op(); |
get_next_xid(), |
op->_state |= ASYNC_OPSTATE_UNKNOWN; |
0, |
op->_flags |= ASYNC_OPFLAGS_SINGLE | ASYNC_OPFLAGS_NORMAL; |
|
|
|
UpdateCimService *msg = new UpdateCimService(get_next_xid(), |
|
op, |
|
true, | true, |
_queueId, | _queueId, |
_capabilities, | _capabilities, |
_mask); | _mask); |
Boolean registered = false; | Boolean registered = false; |
| |
|
|
AsyncMessage *reply = SendWait(msg); | AsyncMessage *reply = SendWait(msg); |
if (reply) | if (reply) |
{ | { |
|
|
if(reply->getMask() & message_mask::ha_reply) | if(reply->getMask() & message_mask::ha_reply) |
{ | { |
if(static_cast<AsyncReply *>(reply)->result == async_results::OK) | if(static_cast<AsyncReply *>(reply)->result == async_results::OK) |
|
{ |
registered = true; | registered = true; |
} | } |
} | } |
|
} |
delete reply; | delete reply; |
} | } |
|
delete msg; |
return registered; | return registered; |
} | } |
| |
| |
Boolean MessageQueueService::deregister_service(void) |
Boolean MessageQueueService::deregister_service() |
{ | { |
| |
// _meta_dispatcher->deregister_module(_queueId); |
_meta_dispatcher->deregister_module(_queueId); |
// return true; |
return true; |
|
|
|
|
AsyncOpNode *op = _meta_dispatcher->get_cached_op(); |
|
op->_state |= ASYNC_OPSTATE_UNKNOWN; |
|
op->_flags |= ASYNC_OPFLAGS_SINGLE | ASYNC_OPFLAGS_NORMAL; |
|
|
|
DeRegisterCimService *msg = new DeRegisterCimService(get_next_xid(), |
|
op, |
|
true, |
|
_queueId); |
|
Boolean deregistered = false; |
|
|
|
|
|
|
|
return _meta_dispatcher->accept_async(static_cast<Message *>(msg)); |
|
} | } |
| |
| |
void MessageQueueService::find_services(String name, |
void MessageQueueService::find_services( |
|
String name, |
Uint32 capabilities, | Uint32 capabilities, |
Uint32 mask, | Uint32 mask, |
Array<Uint32> *results) | Array<Uint32> *results) |
{ | { |
|
|
if( results == 0 ) | if( results == 0 ) |
|
{ |
throw NullPointer(); | throw NullPointer(); |
|
} |
| |
AsyncOpNode *op = get_op(); |
|
results->clear(); | results->clear(); |
| |
FindServiceQueue *req = |
FindServiceQueue *req = new FindServiceQueue( |
new FindServiceQueue(get_next_xid(), |
get_next_xid(), |
op, |
0, |
_queueId, | _queueId, |
true, | true, |
name, | name, |
capabilities, | capabilities, |
mask); | mask); |
| |
|
req->dest = CIMOM_Q_ID; |
|
|
AsyncMessage *reply = SendWait(req); | AsyncMessage *reply = SendWait(req); |
if(reply) | if(reply) |
{ | { |
|
|
} | } |
delete reply; | delete reply; |
} | } |
|
delete req; |
return ; | return ; |
} | } |
| |
void MessageQueueService::enumerate_service(Uint32 queue, message_module *result) | void MessageQueueService::enumerate_service(Uint32 queue, message_module *result) |
{ | { |
if(result == 0) | if(result == 0) |
|
{ |
throw NullPointer(); | throw NullPointer(); |
|
} |
| |
AsyncOpNode *op = get_op(); |
EnumerateService *req = new EnumerateService( |
|
get_next_xid(), |
EnumerateService *req |
0, |
= new EnumerateService(get_next_xid(), |
|
op, |
|
_queueId, | _queueId, |
true, | true, |
queue); | queue); |
|
|
} | } |
delete reply; | delete reply; |
} | } |
|
delete req; |
|
|
return; | return; |
} | } |
| |
Uint32 MessageQueueService::get_next_xid(void) |
Uint32 MessageQueueService::get_next_xid() |
{ | { |
|
static Mutex _monitor; |
|
Uint32 value; |
|
AutoMutex autoMut(_monitor); |
_xid++; | _xid++; |
return _xid.value(); |
value = _xid.value(); |
|
return value; |
|
|
} | } |
| |
PEGASUS_NAMESPACE_END | PEGASUS_NAMESPACE_END |