How to use flatbuffers in python the right way?

The design of FlatBuffers heavily favors languages like C/C++/Rust in attaining maximum speed. The Python implementation mimics what these languages do, but it is very unnatural for Python, so it is not the fastest possible serializer design that you would get if you designed purely for Python.

I haven't benchmarked anything on Python, but a Python specific design would certainly beat FlatBuffers-Python in many cases. One case where the FlatBuffers design will win even in Python is for large files that are accessed sparsely or randomly, since it doesn't actually unpack all the data at once.

You typically use FlatBuffers because you have the performance critical part of your stack in a faster language, and then you also want to be able to process the data in Python elsewhere. If you work purely in Python however, FlatBuffers is possibly not your best pick (unless, again, you work with large sparse data).

Better of course is to not do your heavy lifting in Python in the first place.


  1. You do not refer to any specific link. I guess the performance of flatbuffers is going to be dependent on the serialization from Python while calling the API. Python is known to be slower than, say, C or C++ in that.

  2. Regarding zero-copy - Google (and Wikipedia) is your friend.

  3. Tutorial says "depending on language". What you are saying suggests that in Python you won't get exceptions.

  4. What the documentation say? Do your experiments confirm it? (show us some effort of solving the problem)

  5. Hard to say. What have you tried and what results have you got?


I now made a benchmark in python to compare JSON and flatbuffers and think the answer could benefit someone, so here we go:

The setup is as follows: We got a client server architecture (on the same machine), both in python with sockets and asyncio. The test data is a big dictionary with values as strings, numbers and lists that contain other dictionaries also with string, number and list values. This tree gets max 3 levels deep with around 100 objects per list.

The flatbuffer schema uses tables for the dicts, vectors for the lists and structs for dicts that only use float and int fields.

The test data for the flatbuffer test is:

  • filled into a flatbuffer builder and returned as bytearray (serialization)
  • sent to the server via a socket and asyncio reader/writer
  • transformed from bytearray back into a fb-object and a few fields are accessed by the server (deserialization)
  • the deserialization-time is than sent back to the client.

The test data for the JSON test is:

  • converted to a string via dumps() and than transformed into a bytearray
  • sent to the server via a socket and asyncio reader/writer
  • transformed from bytearray into a string and than via json.loads() back into a dictionary; than the same few fields are accessed by the server (deserialization)
  • the deserialization-time is than sent back to the client.

I know, that there are some points one could argue about, regarding the setup. For example, not transforming the data back into a dict in the flatbuffer test. If someone is actually interested in this I could advance on this test.

But now on to the results:

--- flatbuffers  ---
roundtrip time:   7.010654926300049
serialize time:   6.960820913314819
deserialize time: 0.0
size in byte:     6.365.432
---     json     ---
roundtrip time:   0.7860651016235352
serialize time:   0.3211710453033447
deserialize time: 0.28783655166625977
size in byte:     13.946.172

My conclusion is, that one should not use flatbuffers in python, if you want to edit or create the data fast. There is no way to mutate the data in python, meaning that you would have to rebuilt the flatbuffer every time something changes and this is very slow.

On the bright side, it is very fast to read the data and the byte size is very low compared to JSON. So if you have static data that you want to send or read many times, flatbuffers would be the solution.