Rendered at 14:58:29 GMT+0000 (Coordinated Universal Time) with Cloudflare Workers.
stevemk14ebr 3 days ago [-]
Testing on 5GB of data fully resident in ram is a terrible comparison. Things get hard when you're in the hundreds of gigabytes or more.
MBkkt 5 hours ago [-]
TLDR It's not our choice, but it's meaningful. Because this 5GB is single data segment and literally what you will have in Elastic/etc when you have overall TBs of data. See https://www.elastic.co/docs/deploy-manage/production-guidanc... (single shard is one Lucene index that contains multiple data segments)
malandin 2 days ago [-]
Thanks a lot for your comment! We agree that a dataset as small as 5 GB may sound strange but it was a conscious decision. Check out our blog post to read more about the methodology of this benchmark itself.
Great results! Refreshing to see a project that actually went the extra mile and built the core search engine in C++ from scratch, unlike most similar projects that just wrap an existing library.
malandin 3 days ago [-]
Thank you for your feedback! We have long thought that the space was missing an alternative.
akorchak 3 days ago [-]
Very impressive numbers!
I def would love now to learn more about the algo behind these results
gnusi 3 days ago [-]
Thanks! There some articles on our blog. Check it out!
leonid_y 3 days ago [-]
I can't wait to apply this base to my real-time AI forecasting solutions!
malandin 3 days ago [-]
Hi Leonid! More than happy to talk about this! Find me on social networks by by nickname.
https://blog.serenedb.com/search-benchmark-game-overview