> Direct I/O means no more fsync: no more complexity via background flushes and optimal scheduling of syncs. There's no kernel overhead from copying and coalescing. It essentially provides the performance, control, and simplicity of issuing raw 1:1 I/O requests.
Not true, you still need fsync in direct I/O to ensure durability in power loss situations. Some drives have write caches that means acknowledged writes live in non-volatile memory. So maybe the perf is wildly better because you’re sacrificing durability?
/// Even when using direct I/O, `fsync` is still necessary, as it ensures the device itself has flushed any internal caches.
async fn sync(&self) {
let (fut, fut_ctl) = SignalFuture::new();
self.sender.send(Request::Sync { res: fut_ctl }).unwrap();
fut.await
}
That’s a lot of work creating a whole system that stores data on a raw block device. It would be nice to see this compared to… a filesystem. XFS, ZFS and btrfs are pretty popular.
> Despite serving from same-region datacenters 2 ms from the user, S3 would take 30-200 ms to respond to each request.
200ms seems fairly reasonable to me once we factor in all of the other aspects of S3. A lot of machines would have to die at Amazon for your data to become at risk.
When you have a service and really care about shoving of S3 latencies in the millisecond range, then you propably have enough users that all the tiny images are cached @ edge anyways.
It's a bit weird to present it as an alternative to S3 when it looks like a persistent cache or k/v store. A benchmark against Redis would have been nice for example.
The benchmark for rocks DB is also questionable as the performance depends a lot on how you configure it, and the article's claim that it doesn't support range read doesn't give me confidence in the results.
Also for the descried issue of small images for a frontend, nobody would serve directly from S3 without a caching layer on top.
It's a interesting read for fun, but I am not sure what it solves in the end.
Not true, you still need fsync in direct I/O to ensure durability in power loss situations. Some drives have write caches that means acknowledged writes live in non-volatile memory. So maybe the perf is wildly better because you’re sacrificing durability?