version 1.76, 2003/10/10 14:31:27
|
version 1.88.2.6, 2005/08/23 22:34:37
|
|
|
//%////-*-c++-*-//////////////////////////////////////////////////////////////// |
//%2005//////////////////////////////////////////////////////////////////////// |
// | // |
// Copyright (c) 2000, 2001, 2002 BMC Software, Hewlett-Packard Company, IBM, |
// Copyright (c) 2000, 2001, 2002 BMC Software; Hewlett-Packard Development |
// The Open Group, Tivoli Systems |
// Company, L.P.; IBM Corp.; The Open Group; Tivoli Systems. |
|
// Copyright (c) 2003 BMC Software; Hewlett-Packard Development Company, L.P.; |
|
// IBM Corp.; EMC Corporation, The Open Group. |
|
// Copyright (c) 2004 BMC Software; Hewlett-Packard Development Company, L.P.; |
|
// IBM Corp.; EMC Corporation; VERITAS Software Corporation; The Open Group. |
|
// Copyright (c) 2005 Hewlett-Packard Development Company, L.P.; IBM Corp.; |
|
// EMC Corporation; VERITAS Software Corporation; The Open Group. |
// | // |
// Permission is hereby granted, free of charge, to any person obtaining a copy | // Permission is hereby granted, free of charge, to any person obtaining a copy |
// of this software and associated documentation files (the "Software"), to | // of this software and associated documentation files (the "Software"), to |
|
|
// Author: Mike Day (mdday@us.ibm.com) | // Author: Mike Day (mdday@us.ibm.com) |
// | // |
// Modified By: | // Modified By: |
|
// Amit K Arora, IBM (amita@in.ibm.com) for Bug#1090 |
// | // |
//%///////////////////////////////////////////////////////////////////////////// | //%///////////////////////////////////////////////////////////////////////////// |
| |
|
|
return _thread_pool; | return _thread_pool; |
} | } |
| |
void unload_idle_providers(void); |
// |
|
// MAX_THREADS_PER_SVC_QUEUE_LIMIT |
|
// |
|
// JR Wunderlich Jun 6, 2005 |
|
// |
|
|
|
#define MAX_THREADS_PER_SVC_QUEUE_LIMIT 5000 |
|
#define MAX_THREADS_PER_SVC_QUEUE_DEFAULT 5 |
|
|
|
Uint32 max_threads_per_svc_queue; |
| |
PEGASUS_THREAD_RETURN PEGASUS_THREAD_CDECL MessageQueueService::kill_idle_threads(void *parm) | PEGASUS_THREAD_RETURN PEGASUS_THREAD_CDECL MessageQueueService::kill_idle_threads(void *parm) |
{ | { |
|
|
| |
} | } |
} | } |
exit_thread((PEGASUS_THREAD_RETURN)dead_threads); |
|
return (PEGASUS_THREAD_RETURN)dead_threads; |
#ifdef PEGASUS_POINTER_64BIT |
|
return (PEGASUS_THREAD_RETURN)(Uint64)dead_threads; |
|
#elif PEGASUS_PLATFORM_AIX_RS_IBMCXX |
|
return (PEGASUS_THREAD_RETURN)(unsigned long)dead_threads; |
|
#else |
|
return (PEGASUS_THREAD_RETURN)(Uint32)dead_threads; |
|
#endif |
} | } |
| |
| |
void MessageQueueService::force_shutdown(Boolean destroy_flag) | void MessageQueueService::force_shutdown(Boolean destroy_flag) |
{ | { |
|
return; |
| |
#ifdef MESSAGEQUEUESERVICE_DEBUG | #ifdef MESSAGEQUEUESERVICE_DEBUG |
//l10n | //l10n |
|
|
"Forcing shutdown of CIMOM Message Router"); | "Forcing shutdown of CIMOM Message Router"); |
PEGASUS_STD(cout) << MessageLoader::getMessage(parms) << PEGASUS_STD(endl); | PEGASUS_STD(cout) << MessageLoader::getMessage(parms) << PEGASUS_STD(endl); |
#endif | #endif |
PEGASUS_STD(cout) << "MessageQueueService::force_shutdown()" << PEGASUS_STD(endl); |
|
|
|
| |
| |
MessageQueueService *svc; | MessageQueueService *svc; |
|
|
| |
_polling_sem.signal(); | _polling_sem.signal(); |
| |
PEGASUS_STD(cout) << "Force shutdown found " << counter << " services" << PEGASUS_STD(endl); |
|
|
|
MessageQueueService::_stop_polling = 1; | MessageQueueService::_stop_polling = 1; |
| |
if(destroy_flag == true) | if(destroy_flag == true) |
|
|
svc = _polling_list.remove_last(); | svc = _polling_list.remove_last(); |
while(svc) | while(svc) |
{ | { |
PEGASUS_STD(cout) << "preparing to delete " << svc->getQueueName() << PEGASUS_STD(endl); |
|
|
|
delete svc; | delete svc; |
svc = _polling_list.remove_last(); | svc = _polling_list.remove_last(); |
} | } |
|
|
break; | break; |
} | } |
| |
|
// The polling_routine thread must hold the lock on the |
|
// _polling_thread list while processing incoming messages. |
|
// This lock is used to give this thread ownership of |
|
// services on the _polling_routine list. |
|
|
|
// This is necessary to avoid confict with other threads |
|
// processing the _polling_list |
|
// (e.g., MessageQueueServer::~MessageQueueService). |
|
|
list->lock(); | list->lock(); |
MessageQueueService *service = list->next(0); | MessageQueueService *service = list->next(0); |
|
ThreadStatus rtn = PEGASUS_THREAD_OK; |
while(service != NULL) | while(service != NULL) |
{ | { |
if(service->_incoming.count() > 0 ) |
if ((service->_incoming.count() > 0) && |
|
(service->_die.value() == 0) && |
|
(service->_threads < max_threads_per_svc_queue)) |
|
{ |
|
// The _threads count is used to track the |
|
// number of active threads that have been allocated |
|
// to process messages for this service. |
|
|
|
// The _threads count MUST be incremented while |
|
// the polling_routine owns the _polling_thread |
|
// lock and has ownership of the service object. |
|
|
|
service->_threads++; |
|
try |
|
{ |
|
rtn = _thread_pool->allocate_and_awaken( |
|
service, _req_proc, &_polling_sem); |
|
} |
|
catch (...) |
|
{ |
|
service->_threads--; |
|
|
|
// allocate_and_awaken should never generate an exception. |
|
PEGASUS_ASSERT(0); |
|
} |
|
// if no more threads available, break from processing loop |
|
if (rtn != PEGASUS_THREAD_OK ) |
{ | { |
_thread_pool->allocate_and_awaken(service, _req_proc); |
service->_threads--; |
|
Logger::put(Logger::STANDARD_LOG, System::CIMSERVER, Logger::TRACE, |
|
"Not enough threads to process this request. Skipping."); |
|
|
|
Tracer::trace(TRC_MESSAGEQUEUESERVICE, Tracer::LEVEL2, |
|
"Could not allocate thread for %s. " \ |
|
"Queue has %d messages waiting and %d threads servicing." \ |
|
"Skipping the service for right now. ", |
|
service->getQueueName(), |
|
service->_incoming.count(), |
|
service->_threads.value()); |
|
|
|
pegasus_yield(); |
|
service = NULL; |
|
} |
} | } |
|
if (service != NULL) |
|
{ |
service = list->next(service); | service = list->next(service); |
} | } |
|
} |
list->unlock(); | list->unlock(); |
|
|
if(_check_idle_flag.value() != 0 ) | if(_check_idle_flag.value() != 0 ) |
{ | { |
_check_idle_flag = 0; | _check_idle_flag = 0; |
Thread th(kill_idle_threads, 0, true); |
|
th.run(); |
// try to do idle thread clean up processing when system is not busy |
|
// if system is busy there may not be a thread available to allocate |
|
// so nothing will be done and that is OK. |
|
|
|
if ( _thread_pool->allocate_and_awaken(service, kill_idle_threads, |
|
&_polling_sem) != PEGASUS_THREAD_OK) |
|
{ |
|
Logger::put(Logger::STANDARD_LOG, System::CIMSERVER, Logger::TRACE, |
|
"Not enough threads to kill idle threads. What an irony."); |
|
|
|
Tracer::trace(TRC_MESSAGEQUEUESERVICE, Tracer::LEVEL2, |
|
"Could not allocate thread to kill idle threads." \ |
|
"Skipping. "); |
|
} |
} | } |
} | } |
myself->exit_self( (PEGASUS_THREAD_RETURN) 1 ); | myself->exit_self( (PEGASUS_THREAD_RETURN) 1 ); |
|
|
| |
_mask(mask), | _mask(mask), |
_die(0), | _die(0), |
|
_threads(0), |
_incoming(true, 0), | _incoming(true, 0), |
_callback(true), | _callback(true), |
_incoming_queue_shutdown(0), | _incoming_queue_shutdown(0), |
|
|
_default_op_timeout.tv_sec = 30; | _default_op_timeout.tv_sec = 30; |
_default_op_timeout.tv_usec = 100; | _default_op_timeout.tv_usec = 100; |
| |
_meta_dispatcher_mutex.lock(pegasus_thread_self()); |
max_threads_per_svc_queue = MAX_THREADS_PER_SVC_QUEUE; |
|
|
|
// if requested threads gt MAX_THREADS_PER_SVC_QUEUE_LIMIT |
|
// then set to MAX_THREADS_PER_SVC_QUEUE_LIMIT |
|
|
|
if (max_threads_per_svc_queue > MAX_THREADS_PER_SVC_QUEUE_LIMIT) |
|
{ |
|
max_threads_per_svc_queue = MAX_THREADS_PER_SVC_QUEUE_LIMIT; |
|
} |
|
|
|
// if requested threads eq 0 (unlimited) |
|
// then set to MAX_THREADS_PER_SVC_QUEUE_LIMIT |
|
|
|
if (max_threads_per_svc_queue == 0) |
|
{ |
|
max_threads_per_svc_queue = MAX_THREADS_PER_SVC_QUEUE_DEFAULT; |
|
} |
|
|
|
// cout << "MAX_THREADS_PER_SVC_QUEUE = " << MAX_THREADS_PER_SVC_QUEUE << endl; |
|
// cout << "max_threads_per_svc_queue set to = " << max_threads_per_svc_queue << endl; |
|
|
|
AutoMutex autoMut(_meta_dispatcher_mutex); |
| |
if( _meta_dispatcher == 0 ) | if( _meta_dispatcher == 0 ) |
{ | { |
|
_stop_polling = 0; |
PEGASUS_ASSERT( _service_count.value() == 0 ); | PEGASUS_ASSERT( _service_count.value() == 0 ); |
_meta_dispatcher = new cimom(); | _meta_dispatcher = new cimom(); |
if (_meta_dispatcher == NULL ) | if (_meta_dispatcher == NULL ) |
{ | { |
_meta_dispatcher_mutex.unlock(); |
|
throw NullPointer(); | throw NullPointer(); |
} | } |
_thread_pool = new ThreadPool(0, "MessageQueueService", 0, 0, | _thread_pool = new ThreadPool(0, "MessageQueueService", 0, 0, |
|
|
_polling_thread = new Thread(polling_routine, | _polling_thread = new Thread(polling_routine, |
reinterpret_cast<void *>(&_polling_list), | reinterpret_cast<void *>(&_polling_list), |
false); | false); |
_polling_thread->run(); |
while (!_polling_thread->run()) |
|
{ |
|
pegasus_yield(); |
|
} |
} | } |
_service_count++; | _service_count++; |
| |
if( false == register_service(name, _capabilities, _mask) ) | if( false == register_service(name, _capabilities, _mask) ) |
{ | { |
_meta_dispatcher_mutex.unlock(); |
|
//l10n | //l10n |
//throw BindFailedException("MessageQueueService Base Unable to register with Meta Dispatcher"); | //throw BindFailedException("MessageQueueService Base Unable to register with Meta Dispatcher"); |
MessageLoaderParms parms("Common.MessageQueueService.UNABLE_TO_REGISTER", | MessageLoaderParms parms("Common.MessageQueueService.UNABLE_TO_REGISTER", |
|
|
| |
_polling_list.insert_last(this); | _polling_list.insert_last(this); |
| |
_meta_dispatcher_mutex.unlock(); |
// _meta_dispatcher_mutex.unlock(); //Bug#1090 |
// _callback_thread.run(); | // _callback_thread.run(); |
| |
// _req_thread.run(); | // _req_thread.run(); |
|
|
{ | { |
_die = 1; | _die = 1; |
| |
|
// The polling_routine locks the _polling_list while |
|
// processing the incoming messages for services on the |
|
// list. Deleting the service from the _polling_list |
|
// prior to processing, avoids synchronization issues |
|
// with the _polling_routine. |
|
|
|
_polling_list.remove(this); |
|
|
|
_callback_ready.signal(); |
|
|
|
// ATTN: The code for closing the _incoming queue |
|
// is not working correctly. In OpenPegasus 2.4, |
|
// execution of the following code is very timing |
|
// dependent. This needs to be fix. |
|
// See Bug 4079 for details. |
if (_incoming_queue_shutdown.value() == 0 ) | if (_incoming_queue_shutdown.value() == 0 ) |
{ | { |
_shutdown_incoming_queue(); | _shutdown_incoming_queue(); |
} | } |
_callback_ready.signal(); |
|
| |
_meta_dispatcher_mutex.lock(pegasus_thread_self()); |
// Wait until all threads processing the messages |
|
// for this service have completed. |
|
|
|
while (_threads.value() > 0) |
|
{ |
|
pegasus_yield(); |
|
} |
|
|
|
{ |
|
AutoMutex autoMut(_meta_dispatcher_mutex); |
_service_count--; | _service_count--; |
if (_service_count.value() == 0 ) | if (_service_count.value() == 0 ) |
{ | { |
|
|
delete _thread_pool; | delete _thread_pool; |
_thread_pool = 0; | _thread_pool = 0; |
} | } |
_meta_dispatcher_mutex.unlock(); |
} // mutex unlocks here |
_polling_list.remove(this); |
|
// Clean up in case there are extra stuff on the queue. | // Clean up in case there are extra stuff on the queue. |
while (_incoming.count()) | while (_incoming.count()) |
{ | { |
|
try |
|
{ |
delete _incoming.remove_first(); | delete _incoming.remove_first(); |
} | } |
|
catch (const ListClosed &e) |
|
{ |
|
// If the list is closed, there is nothing we can do. |
|
break; |
|
} |
|
} |
} | } |
| |
void MessageQueueService::_shutdown_incoming_queue(void) | void MessageQueueService::_shutdown_incoming_queue(void) |
|
|
msg->op->_op_dest = this; | msg->op->_op_dest = this; |
msg->op->_request.insert_first(msg); | msg->op->_request.insert_first(msg); |
| |
|
try |
|
{ |
_incoming.insert_last_wait(msg->op); | _incoming.insert_last_wait(msg->op); |
|
_polling_sem.signal(); |
|
} |
|
catch (const ListClosed &) |
|
{ |
|
// This means the queue has already been shut-down (happens when there |
|
// are two AsyncIoctrl::IO_CLOSE messages generated and one got first |
|
// processed. |
|
delete msg; |
|
} |
|
catch (const Permission &) |
|
{ |
|
delete msg; |
|
} |
} | } |
| |
| |
| |
void MessageQueueService::enqueue(Message *msg) throw(IPCException) |
void MessageQueueService::enqueue(Message *msg) |
{ | { |
PEG_METHOD_ENTER(TRC_MESSAGEQUEUESERVICE, "MessageQueueService::enqueue()"); | PEG_METHOD_ENTER(TRC_MESSAGEQUEUESERVICE, "MessageQueueService::enqueue()"); |
| |
|
|
catch(ListClosed & ) | catch(ListClosed & ) |
{ | { |
operation = 0; | operation = 0; |
|
service->_threads--; |
return(0); | return(0); |
} | } |
if( operation ) | if( operation ) |
|
|
service->_handle_incoming_operation(operation); | service->_handle_incoming_operation(operation); |
} | } |
} | } |
|
service->_threads--; |
return(0); | return(0); |
} | } |
| |
|
|
Boolean MessageQueueService::accept_async(AsyncOpNode *op) | Boolean MessageQueueService::accept_async(AsyncOpNode *op) |
{ | { |
if (_incoming_queue_shutdown.value() > 0 ) | if (_incoming_queue_shutdown.value() > 0 ) |
return false; |
return true; |
| |
// ATTN optimization remove the message checking altogether in the base | // ATTN optimization remove the message checking altogether in the base |
// << Mon Feb 18 14:02:20 2002 mdd >> | // << Mon Feb 18 14:02:20 2002 mdd >> |
|
|
| |
MessageQueueService *service = static_cast<MessageQueueService *>(req->op->_service_ptr); | MessageQueueService *service = static_cast<MessageQueueService *>(req->op->_service_ptr); |
| |
|
#ifdef MESSAGEQUEUESERVICE_DEBUG |
|
PEGASUS_STD(cout) << service->getQueueName() << " Received AsyncIoctl::IO_CLOSE " << PEGASUS_STD(endl); |
|
#endif |
|
|
// respond to this message. this is fire and forget, so we don't need to delete anything. | // respond to this message. this is fire and forget, so we don't need to delete anything. |
// this takes care of two problems that were being found | // this takes care of two problems that were being found |
// << Thu Oct 9 10:52:48 2003 mdd >> | // << Thu Oct 9 10:52:48 2003 mdd >> |
|
|
| |
void MessageQueueService::handle_CimServiceStart(CimServiceStart *req) | void MessageQueueService::handle_CimServiceStart(CimServiceStart *req) |
{ | { |
|
|
|
#ifdef MESSAGEQUEUESERVICE_DEBUG |
|
PEGASUS_STD(cout) << getQueueName() << "received START" << PEGASUS_STD(endl); |
|
#endif |
|
|
// clear the stoped bit and update | // clear the stoped bit and update |
_capabilities &= (~(module_capabilities::stopped)); | _capabilities &= (~(module_capabilities::stopped)); |
_make_response(req, async_results::OK); | _make_response(req, async_results::OK); |
|
|
} | } |
void MessageQueueService::handle_CimServiceStop(CimServiceStop *req) | void MessageQueueService::handle_CimServiceStop(CimServiceStop *req) |
{ | { |
|
#ifdef MESSAGEQUEUESERVICE_DEBUG |
|
PEGASUS_STD(cout) << getQueueName() << "received STOP" << PEGASUS_STD(endl); |
|
#endif |
// set the stopeed bit and update | // set the stopeed bit and update |
_capabilities |= module_capabilities::stopped; | _capabilities |= module_capabilities::stopped; |
_make_response(req, async_results::CIM_STOPPED); | _make_response(req, async_results::CIM_STOPPED); |
|
|
| |
Boolean destroy_op = false; | Boolean destroy_op = false; |
| |
if (request->op == false) |
if (request->op == 0) |
{ | { |
request->op = get_op(); | request->op = get_op(); |
request->op->_request.insert_first(request); | request->op->_request.insert_first(request); |
|
|
| |
Uint32 MessageQueueService::get_next_xid(void) | Uint32 MessageQueueService::get_next_xid(void) |
{ | { |
|
static Mutex _monitor; |
|
Uint32 value; |
|
AutoMutex autoMut(_monitor); |
_xid++; | _xid++; |
return _xid.value(); |
value = _xid.value(); |
|
return value; |
|
|
} | } |
| |
PEGASUS_NAMESPACE_END | PEGASUS_NAMESPACE_END |