And I need to serve huge numbers (~trillions) of single-key point queries out of this dataset, out to many consumers over the Internet, with each fetch having as little per-fetch read latency as possible — ideally, with round-trip times resembling e.g. a Redis GET.
Basically, I want the performance properties that you'd get in a single-node scenario from using LMDB and dedicating all your RAM to OS disk cache — but "in the large", with data being sharded into vnodes and then those vnodes being spread+replicated across an elastically auto-scaled set of compute replicas.
You'd think the answer would be a thing that calls itself something like a "distributed KV store", "serverless NoSQL store", etc.
But it seems that all the big "distributed KV store" products and services — DynamoDB, BigTable, Cassandra, Riak KV, FoundationDB, CockroachDB, ScyllaDB, etc. — are all built to optimize for write throughput in a many-distributed-writers use-case, with little concern for per-read latency. They're all "LevelDB in the large", not "LMDB in the large."
Does HN know of any system that would suit my use-case? Or, if not, any ideas why nobody has built one?