Ellis Michael

39 Followers
41 Following
29 Posts

Brett Saiki on our recent Herbie work with @pavpanchekha and team:
https://uwplse.org/2024/05/09/Herbie-Numerical-Compiler.html

We're rethinking Herbie to be target-platform-aware; i.e., for accuracy-aware codegen. Brett also overviews of Herbie's architecture and history. Thanks to DOE for the support!

UW PLSE | Herbie, the Numerical Compiler

@dan oh my god the photos. Just incredible.
@tobinbaker Well said!
@tobinbaker This approach is most similar to (3) but has the advantage of not relying on a Paxos instance to complete, so it’s actually wait-free and completes in a set number of round trips to a quorum.
Linearizable, Wait-free Reads of Replicated State Machines

@dan you’re allowed to borrow it! Just not mutably!
@vish @dan Comic Serif it is, then!

In today's episode of finishing thesis revisions procrastination, I wrote up my thoughts on an important liveness (hyper-)property of distributed systems.

https://ellismichael.com/blog/2023/01/19/deadlock-freedom/

Deadlock Freedom: Beyond Liveness

@djg I just have a "reply-to-this" label where I would put M2, but if that doesn't scale, maybe you could fire off an email to yourself in the M1 thread with a link to M2?

@marcbrooker Interesting blog post! This reminds me of a problem I came up with about a year ago (but never got around to seriously thinking about).

Namely, in a geodistributed system with geodistributed clients, can erasure coding yield any benefits for the latency experienced by those clients in accessing some piece of data? Specifically, here, I'm neglecting the message handling cost entirely and only considering the network latency.

Here's the problem statement I came up with:

Given a graph G = (V, E), latencies for each edge l: E -> R+, an assignment of readers to nodes r: V -> N, d \in N data chunks, and a threshold t <= d, we want to find the optimal placement of these data chunks. That is, we want a function p: V -> N such that \sum_{v \in V} p(v) = d. The latency experienced by a reader at node v is the (maximum) latency between v and a set of nodes whose total weight under p is at least t (minimized across all such sets, of course). Optimal here could mean average (weighted by r) or worst-case etc.

The non-coded case is when t=1.

The question is, keeping G, l, r, and d/t constant (i.e., storing the same amount of data), can increasing t (and changing p accordingly) yield latency benefits under any of our definitions of optimality?