Skip to content

How Do You Evaluate AI Coding Tools?

thoughts

How do you evaluate these new AI coding tools? Why do you like Cursor, Lovable, Windsurf, Bolt, v0 etc more than others?

Functional features to consider

  • Chat, code completion, or multifile generation capabilities
  • Ability to examine code, tests, terminal and browser errors
  • IDE and tooling ecosystem compatibility
  • Navigation assistance for code edits
  • Custom LLM integration options
  • Support for coding style and component requirements
  • Context addition from documentation, tickets, and telemetry
  • Learning knowledge tracking

Performance metrics

  • Code suggestion quality and longevity in codebase
  • Autonomy level and human-AI intervention requirements
  • Code comprehensibility and debuggability
  • Implementation flexibility versus specification-driven approaches

Or is it all vibe checks?


Originally posted on LinkedIn.