2 hours ago · Tech · 0 comments

I've made every dumb Spark mistake at least once. At production scale — real data, real concurrency, real stakeholders yelling in Slack — "it works" and "it works well" are completely different conversations. So I started writing them down. This is the checklist I wish I had taped to my monitor when I started. Every item comes from a real production screwup — mine or someone else's. Before You Write a Single Line Use the DataFrame / Dataset API, not RDDs. RDDs are lambda-driven — Spark can't see inside them, can't optimize them. DataFrames go through the Catalyst optimizer. You get predicate pushdown, filter reordering, Adaptive Query Execution, and cost-based join reordering for free. The RDD API in MLlib is in maintenance mode. Let it go. Pick the right file format.

No comments yet. Log in to reply on the Fediverse. Comments will appear here.