crucible#1058
and the
motivation for that refactoring. As such, statements in the present tense below
("The Crucible upstairs is implemented as a set of async tasks") refer to the
pre-refactoring architecture; the current architecture is described in the
section titled
Counter-Proposal: One Big Task.The Crucible upstairs is implemented as a set of async tasks. These tasks are mostly static, though a few may be spawned at runtime (notably live-repair). The tasks communicate through a mix of message-passing and shared data behind either synchronous or async locks. Even when message-passing is used, it is often a "doorbell" that wakes a task and tells it to check some mutex-protected data structure.
A BlockOp
request ("job") normally passes through eight tasks:
(Dotted lines are a reminder that the "fast ack" optimization means that writes don’t wait for the Downstairs; instead, those jobs are marked as ackable right away. Of course, reads and flushes must go Downstairs)
These tasks are mostly manipulating data stored in a single
tokio::sync::Mutex<Downstairs>
. A single job locks that mutex many times:
1× in
up_listen
inprocess_new_io
→submit_*
3× + 3× in
cmd_loop
→io_send
The first (3×) lock is amortized if multiple jobs are available at the same time
3× in
process_ds_operation
(called inpm_task
)3× + 1× in
up_ds_listen
to ack work to the guestThe first (3×) lock is amortized if multiple jobs are ackable at the same time
In other words, we have to lock the mutex between 11 and 15 times to process a
single job. The locks in io_send
are particularly troublesome, because
they’re highly contended: all 3× downstairs are trying to send data
simultaneously, and the lock forces them to take turns.
With this lock contention, splitting the downstairs client work between multiple
jobs doesn’t actually buy us anything. Running all three cmd_loop
tasks
together using FuturesUnordered
actually showed 1% faster performance in an
quick (unscientific) benchmark!
What data is in the system?
The Crucible upstairs stores a bunch of state, which we can group into a few different categories:
Singleton data
Upstairs state (
UpstairsState
)List of ackable jobs (stored separately, as an optimization to skip iterating over every job and checking if it is ackable)
Guest work map
Global live-repair (e.g. assigned job IDs)
Per-client data
Client state (
DsState
)Statistics (
IOStateCount
,downstairs_errors
,live_repair_completed/
, etc)aborted Last flush (as a
JobId
)New and skipped jobs
Live-repair data (extents limit)
Per-job data
Various IDs
Actual job work (
IOop
)Whether the job is ackable
Whether the job is a replay
Job data (read response, read response hashes)
Per-job + per-client data
IO state (
IOState
)
I went through and hand-annotated which functions use which data, then wrote a Python script to convert it into a per-task table. Since this table is hand-assembled, it may not be 100% correct, but it’s at least mostly right.
Variables ending in a [i]
indicate that only one client’s data is being
accessed; [*]
indicates that multiple clients' data is used.
|
|
|
|
| |
| R/W | R/W | R/W | — | R/W |
| R/W | R/W | R/W | — | R/W |
| R/W | R/W | R/W | — | R/W |
| — | — | R/W | — | — |
| R/W | R/W | R/W | R/W | R/W |
| — | — | W | — | — |
| R/W | — | — | — | R/W |
| — | — | R/W | R/W | — |
| R/W | R/W | R/W | — | R/W |
| — | — | — | R | — |
| R/W | — | R | R | R/W |
| — | — | R | R | — |
| — | — | — | — | R/W |
| R/W | — | — | — | R/W |
| — | — | R/W | — | — |
| — | — | — | R/W | — |
| R/W | — | — | — | — |
| R/W | R/W | R/W | — | R/W |
| — | — | R/W | R/W | — |
| — | — | — | — | R/W |
| — | — | — | — | R/W |
| R/W | — | — | R/W | — |
| R/W | — | — | — | — |
| R/W | — | — | R | — |
| R | — | — | — | — |
| — | — | — | — | W |
| — | — | R/W | — | — |
| R/W | — | — | — | R/W |
| R/W | — | — | R | R/W |
| R/W | — | — | — | — |
| R/W | R/W | R/W | — | R/W |
| — | R/W | R/W | — | — |
| — | — | R/W | — | — |
| — | — | R | — | — |
| R/W | R | R | — | R/W |
| — | — | R/W | R/W | — |
| R | R | R | R | R |
| R/W | R/W | — | — | R/W |
| — | R/W | — | — | — |
| — | — | — | R/W | — |
| R/W | — | R | R/W | R |
For details on how this table was made, see
data_usage_table.py
.
Inter-task signalling
The diagram above is the happy path of job completion, but there are other possibilities! Here’s a map of every inter-task notification that I could find:
Proposal: Separate client locks
As shown in the list above, many pieces of data are per-client.
I experimentally refactored the Upstairs to put that per-client data into a separate data structure, i.e. the set of locks became
downstairs: Mutex<Downstairs>, // pre-existing, but with much less data
clients: [Mutex<DownstairsClient>; 3], // new!
Many existing functions can run using only a single client lock; most notably
io_send
. Functions that require access to both the Downstairs
and
DownstairsClient
data can lock both mutexes (taking either a single client
lock or all three) and run as before.
This is implemented in the
per_client_locks
branch, and reflects a significant refactoring (+7039, -4425
). It could be
split into smaller PRs; indeed,
crucible#1014
is a
mechanical step in this direction.
A good chunk of the changes are simply plumbing, e.g. updating test.rs
to
accomodate the rearranged data:
integration_tests/src/lib.rs | 2 +
upstairs/src/ack_state.rs | 140 +++
upstairs/src/active_jobs.rs | 107 +--
upstairs/src/control.rs | 36 +-
upstairs/src/lib.rs | 4153 +++++++++++++++++++++++++++++++++++++++++++++------------------------------------
upstairs/src/live_repair.rs | 2182 +++++++++++++++++++++++++++----------------
upstairs/src/test.rs | 4844 ++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++---------------------------------
One of the few tricky parts is maintaining per-client, per-job state. The
Downstairs
frequently checks the state of jobs on all three clients, e.g. to
decide whether the job is ackable. To avoid locking the DownstairsClient
for
each job state, I used an per-job atomic integer; each client sets a distinct
set of bits to indicate the job’s current state, and the Downstairs
reads the
entire value to check state across all three clients.
I see a significant random write performance improvement on this branch for large writes:
|
| |
4K | 28.1 MiB/s | 27.6 MiB/s |
1M | 447 MiB/s | 518 MiB/s |
4M | 435 MiB/s | 553 MiB/s |
This is tested on a 128 GiB encrypted disk, with 64 KiB extents and 4K blocks,
using the following fio
script
[global]
filename=/dev/nvme0n1
iodepth=25
ioengine=aio
time_based
runtime=60
numjobs=1
direct=1
stonewall=1
[randwrite-4K]
bs=4K
rw=randwrite
[randwrite-1M]
bs=1M
rw=randwrite
[randwrite-4M]
bs=4M
rw=randwrite
We later discovered that a significant portion of this improvement is simply
because we stopped holding the locks while doing encryption of write data.
crucible#1019
brings in
that improvement without any other changes; and
crucible#1021
does the
same for decryption.
Counter-proposal: One Big Task
I started this exercise focused on performance, but have left it more concerned about correctness: the tangle of tasks interacting with data leaves me nervous about the possibility of weird corners cases.
(For example, we often use the pattern of (1) lock Upstairs::active
,
(2) lock Downstairs
(3) release Upstairs::active
, but not always;
is that a problem? How do we know whether it’s safe or not?)
Many people have been chatting about the "synchronous state machine core, async
at the edges" pattern for writing async code in Rust, e.g.
[_appendix_andrews_async_essay] and
omicron#4332
.
If we wanted to apply that architecture to our system, the first step would be
to consolidate into 4x tasks which own their data (instead of using shared
locks):
Upstairs::run
: owns theUpstairs
(which owns theDownstairs
), performs the work currently done byup_listen
andup_ds_listen
client_run
: thin wrapper around the TCP connection (to start), equivalent topm_task
+cmd_loop
, but probably doing less work because all the data is over inUpstairs::run
Such a system would run a tight loop around a single (large) select!
, which
selects from many possible event sources:
BlockReq
coming from the guestMessages coming back from clients
Live-repair progress
Control server requests
Timeouts and periodic checks
Repair checking
Periodic DTrace logging
Automatic flushes
Each of these events would be dispatched with full (exclusive) access to all
Upstairs
/ Downstairs
/ DownstairsClient
data, comfortable in the
knowledge that no one else could be mutating it.
Over time, we could migrate more per-client work and data into client_run
as
discussed above. I also expect that we’d want to
move encryption and decryption into separate tasks, e.g. as the boundaries of
the system:
I suspect that this would be an even more significant refactoring, because it changes basically everything about Crucible’s architecture. However, I’m beginning to believe that the advantages would be worth the cost: having a synchronous state machine at the core of the Upstairs would make system behavior much easier to reason about, and we could still push heavy workloads (like encryption) into separate tasks (either spawned or persistent).
Determinations
Both proposal and counter-proposal bring us one step towards a future where the client tasks are both independent (not sharing data) and performing meaningful work. The proposal starts by splitting out client data (moving it behind separate locks), then would require future work to implement task-owned data (removing the locks entirely). The counter-proposal starts by switching to owned data (removing the locks), then would require further work to make the client tasks do more work.
The latter seems like a more reasonable order of implementation.
At the 2023-11-13 Storage Huddle, we reached a rough consensus on moving towards the Counter-Proposal described above.
As an incremental step towards that architecture, we discussed making the
6× downstairs tasks (cmd_loop
and pm_task
) "thinner", i.e. not doing any
work using Upstairs
or Downstairs
data. Instead, they will be thin wrappers
around the FramedRead
/ FramedWrite
queues; work will be handled by the
other tasks (presumably doing tx work in up_listen
and rx work in
up_ds_listen
).
Unfortunately, upon further investigation, this incremental step is not
possible. At certain points during normal system operation, the system
relies on having two tasks both touching Upstairs
/ Downstairs
data.
For example, during initial reconciliation, do_reconciliation
(running in the
main upstairs task) blocks that task until it’s complete; it must be fed by
do_reconcile_work
, running in a separate task.
This suggests that the only feasible path forward is a large, non-incremental refactoring.
Post-Determinations
The refactoring was completed in
[crucible#1058
](https://github.com/oxidecomputer/crucible/pull/1058).
It eliminated about 3KLOC from Crucible, and sped up large writes by 20-30%.
Security Considerations
This refactoring does not introduce any new security considerations.
Appendix: Andrew’s async essay
(from #oxide-rust)
I think async makes a lot of things easier, and I have introduced really dumb bugs with hand rolled state machines. It really depends on what type of code you are writing. Many distributed algorithms are designed and documented as state machines in the literature and do not flow linearly like contacting multiple microservices from a client in order. For those protocols, I think you still end up mostly with a hand written state machine and set of callbacks whether intentional or not at the beginning. If you think about it up front like that and sprinkle in the connection handling and I/O as async that lives around the borders of the core, deterministic, testable state machine that is sync and purely logic, you probably get the best of all worlds. This style also eliminates the temptation, mostly, to use cancel-unsafe parts of async libraries like tokio. Ry, myself, and many others have discovered that this way of writing code tends to be more straightforward and much easier to test, when the protocol fits this style, which it takes experience to recognize. The etcd-raft implementation does this to great effect by separating the protocol state machine from all I/O and Go has coroutines and async built in! Other highly lauded systems like FoundationDB and TigerBeetle also build in this style, with TigerBeetle being the most extreme I’ve seen by also not allocating in the core algorithm.
However, because of Cliff’s take on async-io being a state machine transformation, I tried to think of how you could write this type of code to be clearer and more testable, and that’s why I experimented with bog to see if having complete control over the async executor could make this easier and pure async. TBH, I don’t think I succeeded, and that code is nothing more than a rough prototype. It also enforced cancellation and structured concurrency to help make async rust a touch bit more safe, but the conclusion I think I’ve come to is to pick your battles with async and use it judiciously.
For distributed algorithms, write them to be deterministic and message/callback based with an explicit state machine and property based tests them, run them through kani, use some symbolic execution, fuzzing, whatever.. Then make all the IO async and make it as cancel safe as possible. Using the best executor we can, which is tokio currently, and having a library like Rain’s cancel safe one is the best way to ensure that the async code we do write is as safe as it can be within the constraints of rust, which is not going to change. This also allows us to use async where it shines, which is I/O in the first place.
On the other hand if you are hand writing clients or sequential/concurrent operations that don’t require the complexity of a handwritten state machine, then just write your code that way. It will almost certainly be much shorter, more correct, and easier to read than the alternative.
The hard part is when the case is in between, like LRTQ. I wrote it in callback style with async at the edges and all the bugs found after merge were in the async code, but I also didn’t test that code nearly as much for various reasons. The core was rock solid, but it probably took me a lot longer to get there than if I had gone full async. The reason was that the protocol is just not that complex and is mostly clients pulling data after connection setup.
One other point is that writing distributed algorithms with hand written
sync/state machine code can be made easier with a good library. Ry and I have
talked about collaborating on that. In fact I was going to write one, called
noio
for "No I/O" when I got distracted by Cliff’s talk and wrote bog in that
time instead. A few days ago I heard this design pattern actually has a cult
followng called sans io. And it’s doable in any
language.
Of course, most of us aren’t writing distributed algorithms most of the time, but thinking about when we should use async should still be paramount early on when designing something and we should think specifically about why. A think that can make this easier is to go ahead and lay out the data structures and state first and then see how it’s manipulated. If it’s mostly straight line manipulation, do it directly whether that’s async or sync. If there are lots of events driving it you may want the callback based manual thingy with a large select over events / channels. Even if you didn’t intend to do things this way, you may end up with it, and a bunch of async calls mixed in that don’t really have to be there. I think the easiest way to end up in that scenario is using mutexes at all. That’s kind of a digression, but if you don’t use mutexes, your dependencies on who uses data becomes clearer.