Back to Blog

Ranking Every AI Tool I Used in 2026: The Honest Tier List

10 min read

Ranking Every AI Tool I Used in 2026: The Honest Tier List

There are too many AI tools in 2026, and most of them aren't worth your money. Here's my honest tier ranking of every chatbot, coding agent, no-code builder, and content tool I've actually used over the last few months — graded both against direct competitors and at the macro level. S-tier is what you should be using. D-tier is what you should avoid. C-tier is where most of the noise is, and the surprises are in there. Here's the full breakdown.

How I Graded These

Each tool gets evaluated through two lenses:

  1. Versus direct competitors — does Notion beat Obsidian, does GPT beat Claude
  2. Versus the broader AI ecosystem — best-in-class in a dying category isn't S-tier

A tool can dominate its category and still drop a tier because the category itself is being eaten by something bigger. That's the move I'm tracking — not just "what's the best chatbot," but "is buying a chatbot worth it given everything else available?"

Chatbots

ChatGPT — B tier (average user)

Most people started here, and most still are. The models are solid — GPT 5.4 is genuinely good. The chatbot itself feels middle-of-the-road. You won't hate the outputs, you won't love them, and if you're a weak prompter you'll get extremely generic responses. We can all spot ChatGPT writing now. For $20/month it gives you image generation that's surprisingly good, plus the standard chat experience.

Gemini — A tier

If you're spending exactly $20/month on a single AI chatbot and want the best daily driver, Gemini is the answer. Not because Gemini 3.1 is better than GPT 5.4 or Opus 4.7 head-to-head, but because Gemini gives you:

  • Enough usage that you actually use it
  • Nano Banana Pro, the best-in-class image generator
  • Native video generation
  • Native video understanding (give it a YouTube video, a Reel, a TikTok — it analyzes the actual video, which the others can't do at all)

For the average user spending $20/month, Gemini wins.

Claude — C tier or S tier (entirely usage-dependent)

This one splits in half:

$20/month plan: C tier. You'll burn through usage on Opus before you've finished a real project. Claude can't do images or video in chat — only text and code artifacts. For $20/month, you can do better elsewhere.

$100 or $200/month plan: S tier. Opus 4.7 is the best model in the game right now (in my opinion), and Anthropic is the best of the three major labs. If you can afford the higher tier, the experience flips completely.

So the answer to "Should I use Claude as my chatbot?" is "What's your budget?"

Grok — C tier

Worse at coding than the others. Worse image model than ChatGPT or Gemini. Worse video model than Gemini's. Its only real edges are fewer guardrails and good real-time awareness of what's happening on Twitter/X right now. If you're a Twitter power user, B+ tier. For everyone else, C tier and you'd never recommend it.

Coding Agents

Claude Code — S tier

There's a reason it's become the industry standard. It's good. Same caveat as Claude chat: don't show up with a $20/month plan and expect serious work. Bare minimum $100/month. There's a vocal contingent claiming Claude Code has been nerfed and Opus 4.6/4.7 are downgrades — but that's mostly the inevitable noise around any popular tool plus a fair amount of skill issues, not a secret nerf plot. Even at $100/month, you're getting roughly a 90% reduction on what your token usage would cost via the raw API.

Codex — S tier

OpenAI's answer to Claude Code. Major advantage: way more usage per dollar than Claude Code. Plus OpenAI's models are good — for 99% of use cases there's no meaningful difference between GPT 5.4 and Opus 4.6/4.7. So when the question is "Claude Code or Codex?" the honest answer is just pick one. They're equivalents.

Antigravity — A tier

Used to be S tier alongside Claude Code and Codex. Dropped because it pairs less well with Gemini 3.1 than the alternatives pair with their native models. Antigravity with Gemini just isn't as good as Claude Code with Opus or Codex with GPT. Which raises the obvious question: why use the Antigravity wrapper instead of Claude Code or Codex directly? Plus, usage limits have started tightening here too.

OpenClaude / Hermes / Paperclip-style tools — C tier (arguably D)

I'm grouping these together because they share the same flaw: performance theater. Lots of motion, dashboards, heartbeats, talking interfaces — but very little forward progress. After months of asking, nobody has shown me a single OpenClaude use case that can't be done inside Claude Code or Codex more efficiently.

If you tell me it's a "morning report" use case, I'm banning you from the channel.

The kicker: Anthropic confirmed you can't use your Max subscription inside OpenClaude — meaning you'd be paying raw API prices for Opus 4.7. There's been a recent docs update where they say "we talked to someone at Anthropic and it's fine" — but I want more proof than a one-liner before I treat that as settled.

Hermes has its self-updating skills feature that gets close to actually being useful, but every 15 tool calls auto-rewriting your skills also flirts with the same "feels like progress" trap. Why not use Claude Code with the skill creator skill and decide for yourself when a skill needs updating?

Open Code — Skipping

I haven't used it enough to grade it intelligently. Reputation is positive. If you're deep in the open-source / open-weight model world, you're probably already past the point of needing my recommendation.

No-Code Builders

Lovable — D tier

Nine months ago, S tier. Today, what's the point? Codex exists. Claude Code exists. Even chatbots can build front ends now. You're paying an insane premium for one-click Supabase setup and one-click deploys, and the alternatives have closed the front-end design gap. When you compare it to the broader ecosystem, it's tough to justify Lovable's existence.

Bolt — D tier

Same problem as Lovable. Felt revolutionary when it launched. Now it's just another version of Lovable at a price you can't justify.

Replit / Base 44 — Probably D

Haven't used them enough to grade definitively, but they're in the same competitive death zone. When you're competing against Claude Code, Codex, and Antigravity — and those tools only get more user-friendly over time — the long-term outlook is grim.

Cursor — B tier

I used Cursor for a long time as my IDE alongside Claude Code. The pitch was: Claude Code does 99% of the work, but Cursor gives you a second set of eyes via GPT 5.4 review. That use case got eaten by the Codex plugin inside Claude Code itself, which now does adversarial code review without leaving the terminal. If you love Composer 2.0 specifically, A tier. For me, it's a great product looking for a sharper differentiator.

Gemini CLI — D tier

Name 10 people who use Gemini CLI who don't work for Google. You can't. Its absence from the conversation tells you everything. Skip it.

n8n — A tier (but niche now)

I was a hard n8n fanboy for a long time. Today, the rules have changed: everything n8n can do, Claude Code can do. Everything n8n can do, Codex can do. That puts the burden on n8n to justify itself.

It still wins for one specific use case: client projects with non-technical teams who need to be hands-on with the automation. The visual UI is great for that. Outside that lane, I can spin up automations faster inside Claude Code than I can inside n8n itself, which is a problem for n8n's long-term viability.

Make / Zapier — D tier

More expensive than n8n. Less flexible. Why bother? I'm not sure who's still picking these up beyond legacy reasons.

Specialty / Content Tools

NotebookLM — S tier

Easily the best Google product right now, and it's free. I've built whole content workflows around NotebookLM + Claude Code via the NotebookLM CLI. It produces outstanding deliverables — slide decks, infographics, podcasts — and it's the single best tool for any research workflow that involves YouTube videos. Use it.

Perplexity — C tier

Used to be S tier. Now it's hard to justify $20/month for "a chatbot that lets you use other chatbots." I still use Perplexity Pro occasionally for fast same-day web fetches with sources — that one use case is sharper than asking ChatGPT or Claude to do a web search. But it's a nice-to-have, not a daily driver.

Claude Design — B tier (great tool, brutal usage)

In a vacuum, S tier. The output quality is a huge step up from baseline Claude Code or any of the front-end design skills out there. The usage limits make it nearly unusable. I'm on the $200/month plan and I get the same Claude Design usage as someone on Pro. That's wild. A simple landing page can eat 5% of weekly usage. A full design system can eat 30%. Until they fix the usage tier structure, B tier is the honest split.

GitHub Copilot — C-/C tier

I literally forgot to mention it during recording. That probably tells you everything. People use it because their company forces them to.

Kling 3.0 — A tier

Solid video model. Direct competition with Veo 3.1 (B tier — getting outdated and pricey).

Seedance 2.0 — A+ tier

If you've seen any standout AI video lately, it's been Seedance. Step up from Veo and Kling. The visuals are wild and the price is surprisingly reasonable.

Nano Banana Pro — S tier

Best-in-class image model and has been for a while.

GPT Image Gen 2 — A tier (provisional)

Just launched. Early plays look very promising — possibly S tier with more time. For now, A tier. The image side is a major differentiator for ChatGPT now that the model is real.

The Real Takeaway

The big players are obvious: Claude Code, Codex, Gemini, Notebook LM. Stay invested there. Stay away from C tier and below — Hermes, OpenClaude, Lovable, Bolt, Make, Zapier — unless there's a specific case you can't get out of.

But here's the actual point: getting good at Claude Code, Codex, or Antigravity isn't really getting good at those tools. It's getting good at AI fundamentals — prompting, planning, iteration, tool use, architecture. Those skills transfer. If Claude Code falls off a cliff next week and the new S-tier harness drops, you'll be ready, because the underlying skill is the same.

The platform-specific tools come and go. The fundamentals don't.

Frequently Asked Questions

Why is Claude split between C tier and S tier?

Because the chatbot experience is dramatically different at $20/month versus $100+/month. On the cheap tier, you'll hit usage walls fast and Claude can't do images or video. On the higher tiers, Opus 4.7 is the best model in the game and the experience is best-in-class. Same product, two different tools depending on your plan.

What about open-source coding agents like Open Code?

I haven't used them enough to grade fairly. Reputation is positive in the open-source community. If you're already running open-weight models on your own hardware, you're past the point where my recommendation matters.

Is the Claude Code Masterclass worth it if I'm new?

If you're trying to go from zero to AI dev and you don't have a technical background, the Masterclass is built specifically for that path. It updates weekly and recently added the agentic OS modules — Claude Code as engine, Obsidian for memory, GWS CLI for Google Workspace. Available inside Chase AI+.

Why is n8n only A tier when it used to be S tier?

Because the alternatives caught up. Everything n8n does, Claude Code or Codex can do — usually faster. The one place n8n still wins is non-technical client teams who need a visual editor for ongoing tweaks. Outside that lane, the differentiator has eroded.

What's the cheapest reasonable AI stack for an individual creator?

Gemini for daily chat (best value at $20/month given the image and video features), Claude Code on the $100 plan if you're building, NotebookLM for research, and pick one image/video model when you need it (Nano Banana Pro for images, Seedance 2.0 for video). That covers most real workflows without burning $500/month on tools.


If you want to go deeper into the tools that actually move the needle, join the free Chase AI community for templates, prompts, and live breakdowns. And if you're serious about building with AI, check out the paid community, Chase AI+, for hands-on guidance on how to make money with AI.