version 1.117, 2005/10/14 17:40:18
|
version 1.123.2.2, 2006/07/28 20:46:41
|
|
|
//%2005//////////////////////////////////////////////////////////////////////// |
//%2006//////////////////////////////////////////////////////////////////////// |
// | // |
// Copyright (c) 2000, 2001, 2002 BMC Software; Hewlett-Packard Development | // Copyright (c) 2000, 2001, 2002 BMC Software; Hewlett-Packard Development |
// Company, L.P.; IBM Corp.; The Open Group; Tivoli Systems. | // Company, L.P.; IBM Corp.; The Open Group; Tivoli Systems. |
|
|
// IBM Corp.; EMC Corporation; VERITAS Software Corporation; The Open Group. | // IBM Corp.; EMC Corporation; VERITAS Software Corporation; The Open Group. |
// Copyright (c) 2005 Hewlett-Packard Development Company, L.P.; IBM Corp.; | // Copyright (c) 2005 Hewlett-Packard Development Company, L.P.; IBM Corp.; |
// EMC Corporation; VERITAS Software Corporation; The Open Group. | // EMC Corporation; VERITAS Software Corporation; The Open Group. |
|
// Copyright (c) 2006 Hewlett-Packard Development Company, L.P.; IBM Corp.; |
|
// EMC Corporation; Symantec Corporation; The Open Group. |
// | // |
// Permission is hereby granted, free of charge, to any person obtaining a copy | // Permission is hereby granted, free of charge, to any person obtaining a copy |
// of this software and associated documentation files (the "Software"), to | // of this software and associated documentation files (the "Software"), to |
|
|
PEGASUS_NAMESPACE_BEGIN | PEGASUS_NAMESPACE_BEGIN |
| |
cimom *MessageQueueService::_meta_dispatcher = 0; | cimom *MessageQueueService::_meta_dispatcher = 0; |
AtomicInt MessageQueueService::_service_count = 0; |
AtomicInt MessageQueueService::_service_count(0); |
AtomicInt MessageQueueService::_xid(1); |
|
Mutex MessageQueueService::_meta_dispatcher_mutex; | Mutex MessageQueueService::_meta_dispatcher_mutex; |
| |
static struct timeval deallocateWait = {300, 0}; | static struct timeval deallocateWait = {300, 0}; |
| |
ThreadPool *MessageQueueService::_thread_pool = 0; | ThreadPool *MessageQueueService::_thread_pool = 0; |
| |
DQueue<MessageQueueService> MessageQueueService::_polling_list(true); |
MessageQueueService::PollingList* MessageQueueService::_polling_list; |
|
Mutex MessageQueueService::_polling_list_mutex; |
| |
Thread* MessageQueueService::_polling_thread = 0; | Thread* MessageQueueService::_polling_thread = 0; |
| |
|
|
| |
Uint32 max_threads_per_svc_queue; | Uint32 max_threads_per_svc_queue; |
| |
PEGASUS_THREAD_RETURN PEGASUS_THREAD_CDECL |
ThreadReturnType PEGASUS_THREAD_CDECL |
MessageQueueService::kill_idle_threads(void *parm) | MessageQueueService::kill_idle_threads(void *parm) |
{ | { |
| |
static struct timeval now, last = {0,0}; | static struct timeval now, last = {0,0}; |
gettimeofday(&now, NULL); |
Time::gettimeofday(&now); |
int dead_threads = 0; | int dead_threads = 0; |
| |
if (now.tv_sec - last.tv_sec > 120) | if (now.tv_sec - last.tv_sec > 120) |
{ | { |
gettimeofday(&last, NULL); |
Time::gettimeofday(&last); |
try | try |
{ | { |
dead_threads = MessageQueueService::_thread_pool->cleanupIdleThreads(); | dead_threads = MessageQueueService::_thread_pool->cleanupIdleThreads(); |
|
|
} | } |
| |
#ifdef PEGASUS_POINTER_64BIT | #ifdef PEGASUS_POINTER_64BIT |
return (PEGASUS_THREAD_RETURN)(Uint64)dead_threads; |
return (ThreadReturnType)(Uint64)dead_threads; |
#elif PEGASUS_PLATFORM_AIX_RS_IBMCXX | #elif PEGASUS_PLATFORM_AIX_RS_IBMCXX |
return (PEGASUS_THREAD_RETURN)(unsigned long)dead_threads; |
return (ThreadReturnType)(unsigned long)dead_threads; |
#else | #else |
return (PEGASUS_THREAD_RETURN)(Uint32)dead_threads; |
return (ThreadReturnType)(Uint32)dead_threads; |
#endif | #endif |
} | } |
| |
PEGASUS_THREAD_RETURN PEGASUS_THREAD_CDECL MessageQueueService::polling_routine(void *parm) |
ThreadReturnType PEGASUS_THREAD_CDECL MessageQueueService::polling_routine(void *parm) |
{ | { |
Thread *myself = reinterpret_cast<Thread *>(parm); | Thread *myself = reinterpret_cast<Thread *>(parm); |
DQueue<MessageQueueService> *list = reinterpret_cast<DQueue<MessageQueueService> *>(myself->get_parm()); |
List<MessageQueueService, Mutex> *list = |
while (_stop_polling.value() == 0) |
reinterpret_cast<List<MessageQueueService, Mutex>*>(myself->get_parm()); |
|
|
|
while (_stop_polling.get() == 0) |
{ | { |
_polling_sem.wait(); | _polling_sem.wait(); |
| |
if (_stop_polling.value() != 0) |
if (_stop_polling.get() != 0) |
{ | { |
break; | break; |
} | } |
|
|
// (e.g., MessageQueueServer::~MessageQueueService). | // (e.g., MessageQueueServer::~MessageQueueService). |
| |
list->lock(); | list->lock(); |
MessageQueueService *service = list->next(0); |
MessageQueueService *service = list->front(); |
ThreadStatus rtn = PEGASUS_THREAD_OK; | ThreadStatus rtn = PEGASUS_THREAD_OK; |
while (service != NULL) | while (service != NULL) |
{ | { |
if ((service->_incoming.count() > 0) && | if ((service->_incoming.count() > 0) && |
(service->_die.value() == 0) && |
(service->_die.get() == 0) && |
(service->_threads < max_threads_per_svc_queue)) |
(service->_threads.get() < max_threads_per_svc_queue)) |
{ | { |
// The _threads count is used to track the | // The _threads count is used to track the |
// number of active threads that have been allocated | // number of active threads that have been allocated |
|
|
"Skipping the service for right now. ", | "Skipping the service for right now. ", |
service->getQueueName(), | service->getQueueName(), |
service->_incoming.count(), | service->_incoming.count(), |
service->_threads.value()); |
service->_threads.get()); |
| |
pegasus_yield(); |
Threads::yield(); |
service = NULL; | service = NULL; |
} | } |
} | } |
if (service != NULL) | if (service != NULL) |
{ | { |
service = list->next(service); |
service = list->next_of(service); |
} | } |
} | } |
list->unlock(); | list->unlock(); |
| |
if (_check_idle_flag.value() != 0) |
if (_check_idle_flag.get() != 0) |
{ | { |
_check_idle_flag = 0; | _check_idle_flag = 0; |
// try to do idle thread clean up processing when system is not busy | // try to do idle thread clean up processing when system is not busy |
|
|
| |
} | } |
} | } |
myself->exit_self( (PEGASUS_THREAD_RETURN) 1 ); |
myself->exit_self( (ThreadReturnType) 1 ); |
return(0); | return(0); |
} | } |
| |
|
|
_mask(mask), | _mask(mask), |
_die(0), | _die(0), |
_threads(0), | _threads(0), |
_incoming(true, 0), |
_incoming(), |
_incoming_queue_shutdown(0) | _incoming_queue_shutdown(0) |
{ | { |
| |
|
|
if (_meta_dispatcher == 0) | if (_meta_dispatcher == 0) |
{ | { |
_stop_polling = 0; | _stop_polling = 0; |
PEGASUS_ASSERT(_service_count.value() == 0); |
PEGASUS_ASSERT(_service_count.get() == 0); |
_meta_dispatcher = new cimom(); | _meta_dispatcher = new cimom(); |
if (_meta_dispatcher == NULL) | if (_meta_dispatcher == NULL) |
{ | { |
|
|
throw BindFailedException(parms); | throw BindFailedException(parms); |
} | } |
| |
_polling_list.insert_last(this); |
_get_polling_list()->insert_back(this); |
| |
} | } |
| |
|
|
// prior to processing, avoids synchronization issues | // prior to processing, avoids synchronization issues |
// with the _polling_routine. | // with the _polling_routine. |
| |
_polling_list.remove(this); |
// ATTN: added to prevent assertion in List in which the list does not |
|
// contain this element. |
|
|
|
if (_get_polling_list()->contains(this)) |
|
_get_polling_list()->remove(this); |
| |
// ATTN: The code for closing the _incoming queue | // ATTN: The code for closing the _incoming queue |
// is not working correctly. In OpenPegasus 2.5, | // is not working correctly. In OpenPegasus 2.5, |
// execution of the following code is very timing | // execution of the following code is very timing |
// dependent. This needs to be fix. | // dependent. This needs to be fix. |
// See Bug 4079 for details. | // See Bug 4079 for details. |
if (_incoming_queue_shutdown.value() == 0) |
if (_incoming_queue_shutdown.get() == 0) |
{ | { |
_shutdown_incoming_queue(); | _shutdown_incoming_queue(); |
} | } |
|
|
// Wait until all threads processing the messages | // Wait until all threads processing the messages |
// for this service have completed. | // for this service have completed. |
| |
while (_threads.value() > 0) |
while (_threads.get() > 0) |
{ | { |
pegasus_yield(); |
Threads::yield(); |
} | } |
| |
{ | { |
AutoMutex autoMut(_meta_dispatcher_mutex); | AutoMutex autoMut(_meta_dispatcher_mutex); |
_service_count--; | _service_count--; |
if (_service_count.value() == 0) |
if (_service_count.get() == 0) |
{ | { |
| |
_stop_polling++; | _stop_polling++; |
|
|
while (_incoming.count()) | while (_incoming.count()) |
{ | { |
try { | try { |
delete _incoming.remove_first(); |
delete _incoming.dequeue(); |
} catch (const ListClosed &e) | } catch (const ListClosed &e) |
{ | { |
// If the list is closed, there is nothing we can do. | // If the list is closed, there is nothing we can do. |
|
|
| |
void MessageQueueService::_shutdown_incoming_queue() | void MessageQueueService::_shutdown_incoming_queue() |
{ | { |
if (_incoming_queue_shutdown.value() > 0) |
if (_incoming_queue_shutdown.get() > 0) |
return; | return; |
| |
AsyncIoctl *msg = new AsyncIoctl( | AsyncIoctl *msg = new AsyncIoctl( |
get_next_xid(), |
|
0, | 0, |
_queueId, | _queueId, |
_queueId, | _queueId, |
|
|
msg->op->_state &= ~ASYNC_OPSTATE_COMPLETE; | msg->op->_state &= ~ASYNC_OPSTATE_COMPLETE; |
| |
msg->op->_op_dest = this; | msg->op->_op_dest = this; |
msg->op->_request.insert_first(msg); |
msg->op->_request.reset(msg); |
try { | try { |
_incoming.insert_last_wait(msg->op); |
_incoming.enqueue_wait(msg->op); |
_polling_sem.signal(); | _polling_sem.signal(); |
} catch (const ListClosed &) | } catch (const ListClosed &) |
{ | { |
|
|
} | } |
| |
| |
PEGASUS_THREAD_RETURN PEGASUS_THREAD_CDECL MessageQueueService::_req_proc( |
ThreadReturnType PEGASUS_THREAD_CDECL MessageQueueService::_req_proc( |
void * parm) | void * parm) |
{ | { |
MessageQueueService* service = | MessageQueueService* service = |
|
|
try | try |
{ | { |
| |
if (service->_die.value() != 0) |
if (service->_die.get() != 0) |
{ | { |
service->_threads--; | service->_threads--; |
return (0); | return (0); |
|
|
{ | { |
try | try |
{ | { |
operation = service->_incoming.remove_first(); |
operation = service->_incoming.dequeue(); |
} | } |
catch (ListClosed &) | catch (ListClosed &) |
{ | { |
|
|
if (op->_flags & ASYNC_OPFLAGS_SAFE_CALLBACK) | if (op->_flags & ASYNC_OPFLAGS_SAFE_CALLBACK) |
{ | { |
| |
Message *msg = op->get_request(); |
Message *msg = op->removeRequest(); |
if (msg && (msg->getMask() & message_mask::ha_async)) | if (msg && (msg->getMask() & message_mask::ha_async)) |
{ | { |
if (msg->getType() == async_messages::ASYNC_LEGACY_OP_START) | if (msg->getType() == async_messages::ASYNC_LEGACY_OP_START) |
|
|
delete msg; | delete msg; |
} | } |
| |
msg = op->get_response(); |
msg = op->removeResponse(); |
if (msg && (msg->getMask() & message_mask::ha_async)) | if (msg && (msg->getMask() & message_mask::ha_async)) |
{ | { |
if (msg->getType() == async_messages::ASYNC_LEGACY_OP_RESULT) | if (msg->getType() == async_messages::ASYNC_LEGACY_OP_RESULT) |
|
|
// << Tue Feb 19 14:10:38 2002 mdd >> | // << Tue Feb 19 14:10:38 2002 mdd >> |
operation->lock(); | operation->lock(); |
| |
Message *rq = operation->_request.next(0); |
Message *rq = operation->_request.get(); |
| |
// optimization <<< Thu Mar 7 21:04:05 2002 mdd >>> | // optimization <<< Thu Mar 7 21:04:05 2002 mdd >>> |
// move this to the bottom of the loop when the majority of | // move this to the bottom of the loop when the majority of |
|
|
// divert legacy messages to handleEnqueue | // divert legacy messages to handleEnqueue |
if ((rq != 0) && (!(rq->getMask() & message_mask::ha_async))) | if ((rq != 0) && (!(rq->getMask() & message_mask::ha_async))) |
{ | { |
rq = operation->_request.remove_first() ; |
operation->_request.release(); |
operation->unlock(); | operation->unlock(); |
// delete the op node | // delete the op node |
operation->release(); | operation->release(); |
|
|
| |
AsyncLegacyOperationResult *async_result = | AsyncLegacyOperationResult *async_result = |
new AsyncLegacyOperationResult( | new AsyncLegacyOperationResult( |
async->getKey(), |
|
async->getRouting(), |
|
op, | op, |
response); | response); |
_completeAsyncResponse( | _completeAsyncResponse( |
|
|
| |
Boolean MessageQueueService::accept_async(AsyncOpNode *op) | Boolean MessageQueueService::accept_async(AsyncOpNode *op) |
{ | { |
if (_incoming_queue_shutdown.value() > 0) |
if (_incoming_queue_shutdown.get() > 0) |
return false; | return false; |
if (_polling_thread == NULL) | if (_polling_thread == NULL) |
{ | { |
_polling_thread = new Thread( | _polling_thread = new Thread( |
polling_routine, | polling_routine, |
reinterpret_cast<void *>(&_polling_list), |
reinterpret_cast<void *>(_get_polling_list()), |
false); | false); |
ThreadStatus tr = PEGASUS_THREAD_OK; | ThreadStatus tr = PEGASUS_THREAD_OK; |
while ( (tr =_polling_thread->run()) != PEGASUS_THREAD_OK) | while ( (tr =_polling_thread->run()) != PEGASUS_THREAD_OK) |
{ | { |
if (tr == PEGASUS_THREAD_INSUFFICIENT_RESOURCES) | if (tr == PEGASUS_THREAD_INSUFFICIENT_RESOURCES) |
pegasus_yield(); |
Threads::yield(); |
else | else |
throw Exception(MessageLoaderParms("Common.MessageQueueService.NOT_ENOUGH_THREAD", | throw Exception(MessageLoaderParms("Common.MessageQueueService.NOT_ENOUGH_THREAD", |
"Could not allocate thread for the polling thread.")); | "Could not allocate thread for the polling thread.")); |
|
|
// ATTN optimization remove the message checking altogether in the base | // ATTN optimization remove the message checking altogether in the base |
// << Mon Feb 18 14:02:20 2002 mdd >> | // << Mon Feb 18 14:02:20 2002 mdd >> |
op->lock(); | op->lock(); |
Message *rq = op->_request.next(0); |
Message *rq = op->_request.get(); |
Message *rp = op->_response.next(0); |
Message *rp = op->_response.get(); |
op->unlock(); | op->unlock(); |
| |
if ((rq != 0 && (true == messageOK(rq))) || | if ((rq != 0 && (true == messageOK(rq))) || |
(rp != 0 && (true == messageOK(rp))) && _die.value() == 0) |
(rp != 0 && (true == messageOK(rp))) && _die.get() == 0) |
{ | { |
_incoming.insert_last_wait(op); |
_incoming.enqueue_wait(op); |
_polling_sem.signal(); | _polling_sem.signal(); |
return true; | return true; |
} | } |
|
|
| |
Boolean MessageQueueService::messageOK(const Message *msg) | Boolean MessageQueueService::messageOK(const Message *msg) |
{ | { |
if (_incoming_queue_shutdown.value() > 0) |
if (_incoming_queue_shutdown.get() > 0) |
return false; | return false; |
return true; | return true; |
} | } |
|
|
| |
AsyncReply *reply = new AsyncReply( | AsyncReply *reply = new AsyncReply( |
async_messages::HEARTBEAT, | async_messages::HEARTBEAT, |
req->getKey(), |
|
req->getRouting(), |
|
0, | 0, |
req->op, | req->op, |
async_results::OK, | async_results::OK, |
|
|
// ensure we do not accept any further messages | // ensure we do not accept any further messages |
| |
// ensure we don't recurse on IO_CLOSE | // ensure we don't recurse on IO_CLOSE |
if (_incoming_queue_shutdown.value() > 0) |
if (_incoming_queue_shutdown.get() > 0) |
break; | break; |
| |
// set the closing flag | // set the closing flag |
|
|
AsyncOpNode *operation; | AsyncOpNode *operation; |
try | try |
{ | { |
operation = service->_incoming.remove_first(); |
operation = service->_incoming.dequeue(); |
} | } |
catch(IPCException &) | catch(IPCException &) |
{ | { |
|
|
break; | break; |
} // message processing loop | } // message processing loop |
| |
// shutdown the AsyncDQueue |
// shutdown the AsyncQueue |
service->_incoming.shutdown_queue(); |
service->_incoming.close(); |
return; | return; |
} | } |
| |
|
|
| |
void MessageQueueService::return_op(AsyncOpNode *op) | void MessageQueueService::return_op(AsyncOpNode *op) |
{ | { |
PEGASUS_ASSERT(op->read_state() & ASYNC_OPSTATE_RELEASED); |
PEGASUS_ASSERT(op->_state & ASYNC_OPSTATE_RELEASED); |
delete op; | delete op; |
} | } |
| |
|
|
if (!(msg->getMask() & message_mask::ha_async)) | if (!(msg->getMask() & message_mask::ha_async)) |
{ | { |
AsyncLegacyOperationStart *wrapper = new AsyncLegacyOperationStart( | AsyncLegacyOperationStart *wrapper = new AsyncLegacyOperationStart( |
get_next_xid(), |
|
op, | op, |
destination, | destination, |
msg, | msg, |
|
|
} | } |
else | else |
{ | { |
op->_request.insert_first(msg); |
op->_request.reset(msg); |
(static_cast<AsyncMessage *>(msg))->op = op; | (static_cast<AsyncMessage *>(msg))->op = op; |
} | } |
return _meta_dispatcher->route_async(op); | return _meta_dispatcher->route_async(op); |
|
|
if (op == 0) | if (op == 0) |
{ | { |
op = get_op(); | op = get_op(); |
op->_request.insert_first(msg); |
op->_request.reset(msg); |
if (mask & message_mask::ha_async) | if (mask & message_mask::ha_async) |
{ | { |
(static_cast<AsyncMessage *>(msg))->op = op; | (static_cast<AsyncMessage *>(msg))->op = op; |
|
|
if (request->op == 0) | if (request->op == 0) |
{ | { |
request->op = get_op(); | request->op = get_op(); |
request->op->_request.insert_first(request); |
request->op->_request.reset(request); |
destroy_op = true; | destroy_op = true; |
} | } |
| |
|
|
| |
request->op->_client_sem.wait(); | request->op->_client_sem.wait(); |
| |
request->op->lock(); |
AsyncReply* rpl = static_cast<AsyncReply *>(request->op->removeResponse()); |
AsyncReply * rpl = static_cast<AsyncReply *>(request->op->_response.remove_first()); |
|
rpl->op = 0; | rpl->op = 0; |
request->op->unlock(); |
|
| |
if (destroy_op == true) | if (destroy_op == true) |
{ | { |
request->op->lock(); | request->op->lock(); |
request->op->_request.remove(request); |
request->op->_request.release(); |
request->op->_state |= ASYNC_OPSTATE_RELEASED; | request->op->_state |= ASYNC_OPSTATE_RELEASED; |
request->op->unlock(); | request->op->unlock(); |
return_op(request->op); | return_op(request->op); |
|
|
Uint32 mask) | Uint32 mask) |
{ | { |
RegisterCimService *msg = new RegisterCimService( | RegisterCimService *msg = new RegisterCimService( |
get_next_xid(), |
|
0, | 0, |
true, | true, |
name, | name, |
|
|
Boolean MessageQueueService::update_service(Uint32 capabilities, Uint32 mask) | Boolean MessageQueueService::update_service(Uint32 capabilities, Uint32 mask) |
{ | { |
UpdateCimService *msg = new UpdateCimService( | UpdateCimService *msg = new UpdateCimService( |
get_next_xid(), |
|
0, | 0, |
true, | true, |
_queueId, | _queueId, |
|
|
results->clear(); | results->clear(); |
| |
FindServiceQueue *req = new FindServiceQueue( | FindServiceQueue *req = new FindServiceQueue( |
get_next_xid(), |
|
0, | 0, |
_queueId, | _queueId, |
true, | true, |
|
|
} | } |
| |
EnumerateService *req = new EnumerateService( | EnumerateService *req = new EnumerateService( |
get_next_xid(), |
|
0, | 0, |
_queueId, | _queueId, |
true, | true, |
|
|
return; | return; |
} | } |
| |
Uint32 MessageQueueService::get_next_xid() |
MessageQueueService::PollingList* MessageQueueService::_get_polling_list() |
{ | { |
static Mutex _monitor; |
_polling_list_mutex.lock(); |
Uint32 value; |
|
AutoMutex autoMut(_monitor); |
if (!_polling_list) |
_xid++; |
_polling_list = new PollingList; |
value = _xid.value(); |
|
return value; |
_polling_list_mutex.unlock(); |
| |
|
return _polling_list; |
} | } |
| |
PEGASUS_NAMESPACE_END | PEGASUS_NAMESPACE_END |