Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

This isn’t true anymore we are way beyond 2014 Hadoop (what the blog post is about) at this point.

Go try doing an aggregation of 650gb of json data using normal CLI tools vs duckdb or clickhouse. These tools are pipelining and parallelizing in a way that isn’t easy to do with just GNU Parallel (trust me, I’ve tried).



Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: