Join us

Faster Index I/O with NVMe SSDs

A search service (Marginalia Search) gutted its old index internals and dropped memory-mapped B-trees. In their place: a deterministic, block-aligned skip list tuned for direct reads on NVMe SSDs.

It runs on 128KB block sizes, uses custom buffer pools, and leans hard on io_uring for async position lookups. The payoff? Noticeably faster reads and cleaner latency across the board.

Why it matters: More systems are ditching mmap and building SSD-first, hardware-aware data structures. Marginalia just joined the modern camp.


Let's keep in touch!

Stay updated with my latest posts and news. I share insights, updates, and exclusive content.

Unsubscribe anytime. By subscribing, you share your email with @faun and accept our Terms & Privacy.

Give a Pawfive to this post!


Only registered users can post comments. Please, login or signup.

Start writing about what excites you in tech — connect with developers, grow your voice, and get rewarded.

Join other developers and claim your FAUN.dev() account now!

Avatar

The FAUN

@faun
A worldwide community of developers and DevOps enthusiasts!
Developer Influence
3k

Influence

302k

Total Hits

3712

Posts