← Playbook

Guides and how to get started — How the four scores work and why comparisons stay honest across categories.

Scoring

How I score it

Apples to oranges—on purpose.

Vendors want you to compare their checklist to someone else's checklist. That almost always hides the real question: what does this tool cost to own once it is in your stack—not dollars per seat, but time, exceptions, glue, and who has to babysit it when the founder is on a plane.

I use the same four dimensions on every product so I can put a form builder next to a database next to an automation layer and still have a single frame. The scores are not a substitute for the write-up; they are a shared ruler so you can scan, sort, and argue without pretending every tool solves the same job.

Each dimension is intentionally ordinary language. If the label does not land, the review is doing the real work—but if the label lands, you already know where to dig.

The four dimensions at a glance

The same four dimensions, every time

1–5 each · 20 composite

  • Vibe Ready

    Would a non-technical founder reach for it with confidence?

  • Time to Wow

    How fast from signup to something you can show someone?

  • Ease of Use

    Can a PM own it day-to-day without an engineer on call?

  • Depth of Value

    Does it grow with you—or hit a hard ceiling in six months?

Composite (out of 20)

Four scores, each 1–5, summed to 20. The composite is deliberately blunt: a signal, not a verdict. It helps when you are triaging five tools in different categories and need a consistent shape before you read the nuance. It will lie if you treat it as precision—so read the review.

Why "apples to oranges" is the point

Honest comparisons across categories are hard because the products are not interchangeable. Forcing the same four questions anyway does two things: it surfaces where a tool is strong or weak in human terms, and it stops the spreadsheet from pretending a database and a form product are competing on the same row of features. You still compare—but you compare tradeoffs, not fake parity.

Vibe Ready — why it exists

If a founder will not touch it without an engineer in the room, it is not a neutral choice—it is a staffing decision. Vibe Ready captures whether the tool meets people where they are: clear defaults, credible templates, and a path that does not feel embarrassing on day one. Confidence at the keyboard matters as much as raw capability.

Time to Wow — why it exists

Speed to a credible artifact matters. A tool that needs two weeks of setup before it earns a place in a meeting is borrowing runway you might not have. Time to Wow is not "is it powerful"; it is "how fast can you show something real to someone you trust to tell you the truth."

Ease of Use — why it exists

Operability beats brochureware. I score for whether a PM-sized owner can run it week to week—permissions, debugging a bad row, fixing a broken automation—without constantly pulling an engineer off product work. If the only person who can safely change it is a contractor, the cost of ownership just went up.

Depth of Value — why it exists

Some tools are perfect for ninety days and dangerous at twelve months. Depth of Value is where I write the ceiling: where data lives, what happens when volume doubles, what breaks when the workflow gets honest, and whether you are building equity in a system—or renting momentum.