Skip to content

How Do You Evaluate AI Coding Tools?

thoughts 1 min read

How do you evaluate these new AI coding tools? Why do you like Cursor, Lovable, Windsurf, Bolt, v0 etc more than others?

Functional features to consider

  • Chat, code completion, or multifile generation capabilities
  • Ability to examine code, tests, terminal and browser errors
  • IDE and tooling ecosystem compatibility
  • Navigation assistance for code edits
  • Custom LLM integration options
  • Support for coding style and component requirements
  • Context addition from documentation, tickets, and telemetry
  • Learning knowledge tracking

Performance metrics

  • Code suggestion quality and longevity in codebase
  • Autonomy level and human-AI intervention requirements
  • Code comprehensibility and debuggability
  • Implementation flexibility versus specification-driven approaches

Or is it all vibe checks?


Originally posted on LinkedIn.

Navigate with