1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 236 237 238 239 240 241 242 243
|
.. _issue: https://github.com/mosquito/aio-pika/issues
.. _pull request: https://github.com/mosquito/aio-pika/compare
.. _aio-pika: https://github.com/mosquito/aio-pika
.. _official tutorial: https://www.rabbitmq.com/tutorials/tutorial-six-python.html
.. _rpc:
Remote procedure call (RPC)
===========================
.. warning::
This is a beta version of the port from `official tutorial`_. Please when you found an
error create `issue`_ or `pull request`_ for me.
This implementation is a part of official tutorial.
Since version 1.7.0 `aio-pika`_ has :ref:`patterns submodule <patterns-rpc>`.
You might use :class:`aio_pika.patterns.RPC` for real projects.
.. note::
Using the `aio-pika`_ async Python client
.. note::
**Prerequisites**
This tutorial assumes RabbitMQ is installed_ and running on localhost on standard port (`5672`).
In case you use a different host, port or credentials, connections settings would require adjusting.
.. _installed: https://www.rabbitmq.com/download.html
**Where to get help**
If you're having trouble going through this tutorial you can `contact us`_ through the mailing list.
.. _contact us: https://groups.google.com/forum/#!forum/rabbitmq-users
In the :ref:`second tutorial <work-queues>` we learned how to use *Work Queues* to distribute
time-consuming tasks among multiple workers.
But what if we need to run a function on a remote computer and wait for the result? Well, that's a
different story. This pattern is commonly known as *Remote Procedure Call or RPC*.
In this tutorial we're going to use RabbitMQ to build an RPC system: a client and a scalable RPC server.
As we don't have any time-consuming tasks that are worth distributing, we're going to create a dummy
RPC service that returns Fibonacci numbers.
Client interface
++++++++++++++++
To illustrate how an RPC service could be used we're going to create a simple client class. It's going to expose
a method named call which sends an RPC request and blocks until the answer is received:
.. code-block:: python
async def main():
fibonacci_rpc = FibonacciRpcClient()
result = await fibonacci_rpc.call(4)
print("fib(4) is %r" % result)
.. note::
**A note on RPC**
Although RPC is a pretty common pattern in computing, it's often criticised. The
problems arise when a programmer is not aware whether a function call is local or
if it's a slow RPC. Confusions like that result in an unpredictable system and adds
unnecessary complexity to debugging. Instead of simplifying software, misused RPC can
result in unmaintainable spaghetti code.
Bearing that in mind, consider the following advice:
* Make sure it's obvious which function call is local and which is remote.
* Document your system. Make the dependencies between components clear.
* Handle error cases. How should the client react when the RPC server is down for a long time?
When in doubt avoid RPC. If you can, you should use an asynchronous pipeline - instead
of RPC-like blocking, results are asynchronously pushed to a next computation stage.
Callback queue
++++++++++++++
In general doing RPC over RabbitMQ is easy. A client sends a request message and a server
replies with a response message. In order to receive a response the client needs to send
a 'callback' queue address with the request. Let's try it:
.. code-block:: python
async def main():
...
# Queue for results
callback_queue = await channel.declare_queue(exclusive=True)
await channel.default_exchange.publish(
Message(
request,
reply_to=callback_queue.name
),
routing_key='rpc_queue'
)
# ... and some code to read a response message from the callback_queue ...
...
.. note::
**Message properties**
The AMQP protocol predefines a set of 14 properties that go with a message. Most of the
properties are rarely used, with the exception of the following:
* `delivery_mode`: Marks a message as persistent (with a value of 2) or transient (any other value). You may
remember this property from the :ref:`second tutorial <work-queues>`.
* `content_type`: Used to describe the mime-type of the encoding. For example for the
often used JSON encoding it is a good practice to set this property to: application/json.
* `reply_to`: Commonly used to name a callback queue.
* `correlation_id`: Useful to correlate RPC responses with requests.
See additional info in :class:`aio_pika.Message`
Correlation id
++++++++++++++
In the method presented above we suggest creating a callback queue for
every RPC request. That's pretty inefficient, but fortunately there
is a better way - let's create a single callback queue per client.
That raises a new issue, having received a response in that queue
it's not clear to which request the response belongs. That's when the
`correlation_id` property is used. We're going to set it to a unique value
for every request. Later, when we receive a message in the callback queue
we'll look at this property, and based on that we'll be able to match a
response with a request. If we see an unknown `correlation_id` value, we
may safely discard the message - it doesn't belong to our requests.
You may ask, why should we ignore unknown messages in the callback queue,
rather than failing with an error? It's due to a possibility of a race
condition on the server side. Although unlikely, it is possible that the
RPC server will die just after sending us the answer, but before sending an
acknowledgment message for the request. If that happens, the restarted
RPC server will process the request again. That's why on the client we
must handle the duplicate responses gracefully, and the RPC should
ideally be idempotent.
Summary
+++++++
.. image:: /_static/tutorial/python-six.svg
:align: center
Our RPC will work like this:
* When the Client starts up, it creates an anonymous exclusive callback queue.
* For an RPC request, the Client sends a message with two properties: `reply_to`,
which is set to the callback queue and `correlation_id`, which is set to a
unique value for every request.
* The request is sent to an rpc_queue queue.
* The RPC worker (aka: server) is waiting for requests on that queue. When a
request appears, it does the job and sends a message with the result back to the
Client, using the queue from the reply_to field.
* The client waits for data on the callback queue. When a message appears, it
checks the `correlation_id` property. If it matches the value from the
request it returns the response to the application.
Putting it all together
+++++++++++++++++++++++
The code for :download:`rpc_server.py <examples/6-rpc/rpc_server.py>`:
.. literalinclude:: examples/6-rpc/rpc_server.py
:language: python
:linenos:
The server code is rather straightforward:
* (34) As usual we start by establishing the connection and declaring the queue.
* (6) We declare our fibonacci function. It assumes only valid positive integer input.
(Don't expect this one to work for big numbers, it's probably the slowest recursive implementation possible).
* (15) We declare a callback for basic_consume, the core of the RPC server.
It's executed when the request is received. It does the work and sends the response back.
The code for :download:`rpc_client.py <examples/6-rpc/rpc_client.py>`:
.. literalinclude:: examples/6-rpc/rpc_client.py
:language: python
:linenos:
The client code is slightly more involved:
* (15) We establish a connection, channel and declare an exclusive 'callback' queue for replies.
* (22) We subscribe to the 'callback' queue, so that we can receive RPC responses.
* (26) The 'on_response' callback executed on every response is doing a very simple job,
for every response message it checks if the correlation_id is the one we're looking for.
If so, it saves the response in self.response and breaks the consuming loop.
* (30) Next, we define our main call method - it does the actual RPC request.
* (31) In this method, first we generate a unique correlation_id number and save it - the 'on_response' callback
function will use this value to catch the appropriate response.
* (36) Next, we publish the request message, with two properties: reply_to and correlation_id.
And finally we return the response back to the user.
Our RPC service is now ready. We can start the server::
$ python rpc_server.py
[x] Awaiting RPC requests
To request a fibonacci number run the client::
$ python rpc_client.py
[x] Requesting fib(30)
The presented design is not the only possible implementation of
a RPC service, but it has some important advantages:
If the RPC server is too slow, you can scale up by just running another one.
Try running a second rpc_server.py in a new console.
On the client side, the RPC requires sending and receiving only one message.
No synchronous calls like queue_declare are required. As a result the RPC client
needs only one network round trip for a single RPC request.
Our code is still pretty simplistic and doesn't try to solve more
complex (but important) problems, like:
* How should the client react if there are no servers running?
* Should a client have some kind of timeout for the RPC?
* If the server malfunctions and raises an exception, should it be forwarded to the client?
* Protecting against invalid incoming messages (eg checking bounds) before processing.
.. note::
If you want to experiment, you may find the rabbitmq-management plugin useful for viewing the queues.
.. note::
This material was adopted from `official tutorial`_ on **rabbitmq.org**.
|