# vim: tabstop=4 shiftwidth=4 softtabstop=4 # Copyright 2010 United States Government as represented by the # Administrator of the National Aeronautics and Space Administration. # All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. """AMQP-based RPC. Queues have consumers and publishers. No fan-out support yet. """ import json import sys import time import traceback import types import uuid from carrot import connection as carrot_connection from carrot import messaging from eventlet import greenpool from eventlet import pools from eventlet import queue import greenlet from nova import context from nova import exception from nova import fakerabbit from nova import flags from nova import log as logging from nova import utils LOG = logging.getLogger('nova.rpc') FLAGS = flags.FLAGS flags.DEFINE_integer('rpc_thread_pool_size', 1024, 'Size of RPC thread pool') flags.DEFINE_integer('rpc_conn_pool_size', 30, 'Size of RPC connection pool') class Connection(carrot_connection.BrokerConnection): """Connection instance object.""" @classmethod def instance(cls, new=True): """Returns the instance.""" if new or not hasattr(cls, '_instance'): params = dict(hostname=FLAGS.rabbit_host, port=FLAGS.rabbit_port, ssl=FLAGS.rabbit_use_ssl, userid=FLAGS.rabbit_userid, password=FLAGS.rabbit_password, virtual_host=FLAGS.rabbit_virtual_host) if FLAGS.fake_rabbit: params['backend_cls'] = fakerabbit.Backend # NOTE(vish): magic is fun! # pylint: disable=W0142 if new: return cls(**params) else: cls._instance = cls(**params) return cls._instance @classmethod def recreate(cls): """Recreates the connection instance. This is necessary to recover from some network errors/disconnects. """ try: del cls._instance except AttributeError, e: # The _instance stuff is for testing purposes. Usually we don't use # it. So don't freak out if it doesn't exist. pass return cls.instance() class Pool(pools.Pool): """Class that implements a Pool of Connections.""" # TODO(comstud): Timeout connections not used in a while def create(self): LOG.debug('Creating new connection') return Connection.instance(new=True) # Create a ConnectionPool to use for RPC calls. We'll order the # pool as a stack (LIFO), so that we can potentially loop through and # timeout old unused connections at some point ConnectionPool = Pool( max_size=FLAGS.rpc_conn_pool_size, order_as_stack=True) class Consumer(messaging.Consumer): """Consumer base class. Contains methods for connecting the fetch method to async loops. """ def __init__(self, *args, **kwargs): for i in xrange(FLAGS.rabbit_max_retries): if i > 0: time.sleep(FLAGS.rabbit_retry_interval) try: super(Consumer, self).__init__(*args, **kwargs) self.failed_connection = False break except Exception as e: # Catching all because carrot sucks fl_host = FLAGS.rabbit_host fl_port = FLAGS.rabbit_port fl_intv = FLAGS.rabbit_retry_interval LOG.error(_('AMQP server on %(fl_host)s:%(fl_port)d is' ' unreachable: %(e)s. Trying again in %(fl_intv)d' ' seconds.') % locals()) self.failed_connection = True if self.failed_connection: LOG.error(_('Unable to connect to AMQP server ' 'after %d tries. Shutting down.'), FLAGS.rabbit_max_retries) sys.exit(1) def fetch(self, no_ack=None, auto_ack=None, enable_callbacks=False): """Wraps the parent fetch with some logic for failed connection.""" # TODO(vish): the logic for failed connections and logging should be # refactored into some sort of connection manager object try: if self.failed_connection: # NOTE(vish): connection is defined in the parent class, we can # recreate it as long as we create the backend too # pylint: disable=W0201 self.connection = Connection.recreate() self.backend = self.connection.create_backend() self.declare() return super(Consumer, self).fetch(no_ack, auto_ack, enable_callbacks) if self.failed_connection: LOG.error(_('Reconnected to queue')) self.failed_connection = False # NOTE(vish): This is catching all errors because we really don't # want exceptions to be logged 10 times a second if some # persistent failure occurs. except Exception, e: # pylint: disable=W0703 if not self.failed_connection: LOG.exception(_('Failed to fetch message from queue: %s' % e)) self.failed_connection = True def attach_to_eventlet(self): """Only needed for unit tests!""" timer = utils.LoopingCall(self.fetch, enable_callbacks=True) timer.start(0.1) return timer class AdapterConsumer(Consumer): """Calls methods on a proxy object based on method and args.""" def __init__(self, connection=None, topic='broadcast', proxy=None): LOG.debug(_('Initing the Adapter Consumer for %s') % topic) self.proxy = proxy self.pool = greenpool.GreenPool(FLAGS.rpc_thread_pool_size) super(AdapterConsumer, self).__init__(connection=connection, topic=topic) self.register_callback(self.process_data) def process_data(self, message_data, message): """Consumer callback to call a method on a proxy object. Parses the message for validity and fires off a thread to call the proxy object method. Message data should be a dictionary with two keys: method: string representing the method to call args: dictionary of arg: value Example: {'method': 'echo', 'args': {'value': 42}} """ LOG.debug(_('received %s') % message_data) # This will be popped off in _unpack_context msg_id = message_data.get('_msg_id', None) ctxt = _unpack_context(message_data) method = message_data.get('method') args = message_data.get('args', {}) message.ack() if not method: # NOTE(vish): we may not want to ack here, but that means that bad # messages stay in the queue indefinitely, so for now # we just log the message and send an error string # back to the caller LOG.warn(_('no method for message: %s') % message_data) if msg_id: msg_reply(msg_id, _('No method for message: %s') % message_data) return self.pool.spawn_n(self._process_data, msg_id, ctxt, method, args) @exception.wrap_exception() def _process_data(self, msg_id, ctxt, method, args): """Thread that maigcally looks for a method on the proxy object and calls it. """ node_func = getattr(self.proxy, str(method)) node_args = dict((str(k), v) for k, v in args.iteritems()) # NOTE(vish): magic is fun! try: rval = node_func(context=ctxt, **node_args) if msg_id: # Check if the result was a generator if isinstance(rval, types.GeneratorType): for x in rval: msg_reply(msg_id, x, None) else: msg_reply(msg_id, rval, None) # This final None tells multicall that it is done. msg_reply(msg_id, None, None) elif isinstance(rval, types.GeneratorType): # NOTE(vish): this iterates through the generator list(rval) except Exception as e: logging.exception('Exception during message handling') if msg_id: msg_reply(msg_id, None, sys.exc_info()) return class TopicAdapterConsumer(AdapterConsumer): """Consumes messages on a specific topic.""" exchange_type = 'topic' def __init__(self, connection=None, topic='broadcast', proxy=None): self.queue = topic self.routing_key = topic self.exchange = FLAGS.control_exchange self.durable = False super(TopicAdapterConsumer, self).__init__(connection=connection, topic=topic, proxy=proxy) class FanoutAdapterConsumer(AdapterConsumer): """Consumes messages from a fanout exchange.""" exchange_type = 'fanout' def __init__(self, connection=None, topic='broadcast', proxy=None): self.exchange = '%s_fanout' % topic self.routing_key = topic unique = uuid.uuid4().hex self.queue = '%s_fanout_%s' % (topic, unique) self.durable = False # Fanout creates unique queue names, so we should auto-remove # them when done, so they're not left around on restart. # Also, we're the only one that should be consuming. exclusive # implies auto_delete, so we'll just set that.. self.exclusive = True LOG.info(_('Created "%(exchange)s" fanout exchange ' 'with "%(key)s" routing key'), dict(exchange=self.exchange, key=self.routing_key)) super(FanoutAdapterConsumer, self).__init__(connection=connection, topic=topic, proxy=proxy) class ConsumerSet(object): """Groups consumers to listen on together on a single connection.""" def __init__(self, connection, consumer_list): self.consumer_list = set(consumer_list) self.consumer_set = None self.enabled = True self.init(connection) def init(self, conn): if not conn: conn = Connection.instance(new=True) if self.consumer_set: self.consumer_set.close() self.consumer_set = messaging.ConsumerSet(conn) for consumer in self.consumer_list: consumer.connection = conn # consumer.backend is set for us self.consumer_set.add_consumer(consumer) def reconnect(self): self.init(None) def wait(self, limit=None): running = True while running: it = self.consumer_set.iterconsume(limit=limit) if not it: break while True: try: it.next() except StopIteration: return except greenlet.GreenletExit: running = False break except Exception as e: LOG.exception(_("Exception while processing consumer")) self.reconnect() # Break to outer loop break def close(self): self.consumer_set.close() class Publisher(messaging.Publisher): """Publisher base class.""" pass class TopicPublisher(Publisher): """Publishes messages on a specific topic.""" exchange_type = 'topic' def __init__(self, connection=None, topic='broadcast'): self.routing_key = topic self.exchange = FLAGS.control_exchange self.durable = False super(TopicPublisher, self).__init__(connection=connection) class FanoutPublisher(Publisher): """Publishes messages to a fanout exchange.""" exchange_type = 'fanout' def __init__(self, topic, connection=None): self.exchange = '%s_fanout' % topic self.queue = '%s_fanout' % topic self.durable = False self.auto_delete = True LOG.info(_('Creating "%(exchange)s" fanout exchange'), dict(exchange=self.exchange)) super(FanoutPublisher, self).__init__(connection=connection) class DirectConsumer(Consumer): """Consumes messages directly on a channel specified by msg_id.""" exchange_type = 'direct' def __init__(self, connection=None, msg_id=None): self.queue = msg_id self.routing_key = msg_id self.exchange = msg_id self.auto_delete = True self.exclusive = True super(DirectConsumer, self).__init__(connection=connection) class DirectPublisher(Publisher): """Publishes messages directly on a channel specified by msg_id.""" exchange_type = 'direct' def __init__(self, connection=None, msg_id=None): self.routing_key = msg_id self.exchange = msg_id self.auto_delete = True super(DirectPublisher, self).__init__(connection=connection) def msg_reply(msg_id, reply=None, failure=None): """Sends a reply or an error on the channel signified by msg_id. Failure should be a sys.exc_info() tuple. """ if failure: message = str(failure[1]) tb = traceback.format_exception(*failure) LOG.error(_("Returning exception %s to caller"), message) LOG.error(tb) failure = (failure[0].__name__, str(failure[1]), tb) with ConnectionPool.item() as conn: publisher = DirectPublisher(connection=conn, msg_id=msg_id) try: publisher.send({'result': reply, 'failure': failure}) except TypeError: publisher.send( {'result': dict((k, repr(v)) for k, v in reply.__dict__.iteritems()), 'failure': failure}) publisher.close() class RemoteError(exception.Error): """Signifies that a remote class has raised an exception. Containes a string representation of the type of the original exception, the value of the original exception, and the traceback. These are sent to the parent as a joined string so printing the exception contains all of the relevent info. """ def __init__(self, exc_type, value, traceback): self.exc_type = exc_type self.value = value self.traceback = traceback super(RemoteError, self).__init__('%s %s\n%s' % (exc_type, value, traceback)) def _unpack_context(msg): """Unpack context from msg.""" context_dict = {} for key in list(msg.keys()): # NOTE(vish): Some versions of python don't like unicode keys # in kwargs. key = str(key) if key.startswith('_context_'): value = msg.pop(key) context_dict[key[9:]] = value context_dict['msg_id'] = msg.pop('_msg_id', None) LOG.debug(_('unpacked context: %s'), context_dict) return RpcContext.from_dict(context_dict) def _pack_context(msg, context): """Pack context into msg. Values for message keys need to be less than 255 chars, so we pull context out into a bunch of separate keys. If we want to support more arguments in rabbit messages, we may want to do the same for args at some point. """ context_d = dict([('_context_%s' % key, value) for (key, value) in context.to_dict().iteritems()]) msg.update(context_d) class RpcContext(context.RequestContext): def __init__(self, *args, **kwargs): msg_id = kwargs.pop('msg_id', None) self.msg_id = msg_id super(RpcContext, self).__init__(*args, **kwargs) def reply(self, *args, **kwargs): msg_reply(self.msg_id, *args, **kwargs) def multicall(context, topic, msg): """Make a call that returns multiple times.""" LOG.debug(_('Making asynchronous call on %s ...'), topic) msg_id = uuid.uuid4().hex msg.update({'_msg_id': msg_id}) LOG.debug(_('MSG_ID is %s') % (msg_id)) _pack_context(msg, context) con_conn = ConnectionPool.get() consumer = DirectConsumer(connection=con_conn, msg_id=msg_id) wait_msg = MulticallWaiter(consumer) consumer.register_callback(wait_msg) publisher = TopicPublisher(connection=con_conn, topic=topic) publisher.send(msg) publisher.close() return wait_msg class MulticallWaiter(object): def __init__(self, consumer): self._consumer = consumer self._results = queue.Queue() self._closed = False def close(self): self._closed = True self._consumer.close() ConnectionPool.put(self._consumer.connection) def __call__(self, data, message): """Acks message and sets result.""" message.ack() if data['failure']: self._results.put(RemoteError(*data['failure'])) else: self._results.put(data['result']) def __iter__(self): return self.wait() def wait(self): while True: rv = None while rv is None and not self._closed: try: rv = self._consumer.fetch(enable_callbacks=True) except Exception: self.close() raise time.sleep(0.01) result = self._results.get() if isinstance(result, Exception): self.close() raise result if result == None: self.close() raise StopIteration yield result def call(context, topic, msg): """Sends a message on a topic and wait for a response.""" rv = multicall(context, topic, msg) # NOTE(vish): return the last result from the multicall rv = list(rv) if not rv: return return rv[-1] def cast(context, topic, msg): """Sends a message on a topic without waiting for a response.""" LOG.debug(_('Making asynchronous cast on %s...'), topic) _pack_context(msg, context) with ConnectionPool.item() as conn: publisher = TopicPublisher(connection=conn, topic=topic) publisher.send(msg) publisher.close() def fanout_cast(context, topic, msg): """Sends a message on a fanout exchange without waiting for a response.""" LOG.debug(_('Making asynchronous fanout cast...')) _pack_context(msg, context) with ConnectionPool.item() as conn: publisher = FanoutPublisher(topic, connection=conn) publisher.send(msg) publisher.close() def generic_response(message_data, message): """Logs a result and exits.""" LOG.debug(_('response %s'), message_data) message.ack() sys.exit(0) def send_message(topic, message, wait=True): """Sends a message for testing.""" msg_id = uuid.uuid4().hex message.update({'_msg_id': msg_id}) LOG.debug(_('topic is %s'), topic) LOG.debug(_('message %s'), message) if wait: consumer = messaging.Consumer(connection=Connection.instance(), queue=msg_id, exchange=msg_id, auto_delete=True, exchange_type='direct', routing_key=msg_id) consumer.register_callback(generic_response) publisher = messaging.Publisher(connection=Connection.instance(), exchange=FLAGS.control_exchange, durable=False, exchange_type='topic', routing_key=topic) publisher.send(message) publisher.close() if wait: consumer.wait() consumer.close() if __name__ == '__main__': # You can send messages from the command line using # topic and a json string representing a dictionary # for the method send_message(sys.argv[1], json.loads(sys.argv[2]))