Skip to content Skip to footer

Stop Syncing Everything by neilk

11 Comments

  • Post Author
    smitty1e
    Posted April 2, 2025 at 1:46 am
  • Post Author
    hank808
    Posted April 2, 2025 at 2:11 am

    [flagged]

  • Post Author
    mrbluecoat
    Posted April 2, 2025 at 2:15 am

    > Graft should be considered Alpha quality software. Thus, don't use it for production workloads yet.

    Beta ETA?

  • Post Author
    conradev
    Posted April 2, 2025 at 2:19 am

    > After a client pulls a graft, it knows exactly what’s changed. It can use that information to determine precisely which pages are still valid and which pages need to be fetched

    Curious how this compares to Cloud-Backed SQLite’s manifest: https://sqlite.org/cloudsqlite/doc/trunk/www/index.wiki

    It’s similar to your design (sending changed pages), but doesn’t need any compute on the server, which I think is a huge win.

  • Post Author
    brcmthrowaway
    Posted April 2, 2025 at 2:27 am

    [flagged]

  • Post Author
    matlin
    Posted April 2, 2025 at 2:28 am

    My ideal version of this is simple: just define the queries you want (no matter how complex) and the you'll get exactly the data you need to fulfill those queries, no more, no less. And the cherry on top would be to have your queries update automatically with changes both locally and remote in close to real-time.

    That's basically what we're doing with Triplit (https://triplit.dev), be it, not with SQL–which is a plus for most developers.

  • Post Author
    canadiantim
    Posted April 2, 2025 at 2:55 am

    How does this compare with Turso? I know it's mentioned in the article (mainly better support for partial replication and arbitrary schemas), but is there also a deeper architectural departure between the two projects?

    Looks really good, great work!

  • Post Author
    wg0
    Posted April 2, 2025 at 3:22 am

    Seems interesting. A very challenging problem to wrap your head around. Anyone working on this is exactly pushing the field forward.

    I'm thinking to give it a try in one of my React Native apps that face very uncertain connectivity.

  • Post Author
    snickell
    Posted April 2, 2025 at 3:58 am

    This is a really interesting project, and a great read. I learned a lot. I'm falling down the rabbit hole pretty hard reading about the "Leap" algorithm (https://www.usenix.org/system/files/atc20-maruf.pdf) it uses to predict remote memory prefetches.

    It's easy to focus on libgraft's SQLite integration (comparing to turso, etc), but I appreciate that the author approached this as a more general and lower-level distributed storage problem. If it proves robust in practice, I could see this being used for a lot more than just sqlite.

    At the same time, I think "low level general solutions" are often unhinged when they're not guided by concrete experience. The author's experience with sqlsync, and applying graft to sqlite on day one, feels like it gives them standing to take a stab at a general solution. I like the approach they came up with, particularly shifting responsibility for reconciliation to the application/client layer. Because reconciliation lives heavily in tradeoff space, it feels right to require the application to think closely about how they want to do it.

    A lot of the questions here are requesting comparison's to existing SQLite replication systems, the article actually has a great section on this topic at the bottom: https://sqlsync.dev/posts/stop-syncing-everything/#compariso…

  • Post Author
    kiitos
    Posted April 2, 2025 at 4:20 am

    The consistency model doesn't seem to make sense.

    https://github.com/orbitinghail/graft/blob/main/docs/design….

    > Graft clients commit locally and then asynchronously attempt to commit remotely. Because Graft enforces Strict Serializability globally, when two clients concurrently commit based on the same snapshot, one commit will succeed and the other will fail.

    OK, but, the API provides only a single commit operation:

    > commit(VolumeId, ClientId, Snapshot LSN, page_count, segments) Commit changes to a Volume if it is safe to do so. The provided Snapshot LSN is the snapshot the commit was based on. Returns the newly committed Snapshot on success.

    So if a client commits something, and it succeeds, presumably locally, then how should that client discover that the "async" propagation of that commit has failed, and therefore everything it's done on top of that successful local commit needs to be rolled-back?

    This model is kind of conflating multiple, very different, notions of "commit" with each other. Usually "commit" means the committed transaction/state/whatever is guaranteed to be valid. But here it seems like a "local commit" can be invalidated at some arbitrary point in the future, and is something totally different than an "async-validated commit"?

  • Post Author
    mhahn
    Posted April 2, 2025 at 4:28 am

    I looked at using turso embedded replicas for a realtime collaboration project and one downside was that each sync operation was fairly expensive. The minimum payload size is 4KB IIRC because it needs to sync the sqlite frame. Then they charge based on the number of sync operations so it wasn't a good fit for this particular use case.

    I'm curious if the graft solution helps with this. The idea of just being able to ship a sqlite db to a mobile client that you can also mutate from a server is really powerful. I ended up basically building my own syncing engine to sync changes between clients and servers.

Leave a comment

In the Shadows of Innovation”

© 2025 HackTech.info. All Rights Reserved.

Sign Up to Our Newsletter

Be the first to know the latest updates

Whoops, you're not connected to Mailchimp. You need to enter a valid Mailchimp API key.