Menu
Sign In Search Podcasts Charts People & Topics Add Podcast API Blog Pricing
Podcast Image

AI squared: AI explained

Chat GPT vs Gemini

21 Feb 2026

Transcription

Chapter 1: What is the main topic discussed in this episode?

2.191 - 30.392

Welcome back to AI Squared, where two minds explore one intelligent future. I'm Ayush. And I'm Mikkel. Today we're launching a new mini-series called AI Versus. Now this is a championship racket, but for AI tools. Every episode is one head-to-head matchup judged with the same scorecard. We're not doing my favorite model takes. We're going to do categories that maps are really used.

0

30.642 - 45.192

We're going to use writing, coding, image generation, accuracy, value for the cost of the premium models, thinking modes, research capabilities, voice, UI, speed, and privacy.

0

Chapter 2: What is the purpose of the AI Versus mini-series?

46.893 - 79.077

Episode 1 is a big one. ChatGPT GPT 5.2 thinking versus Google Gemini, specifically the top tier Gemini model like Gemini 2.5 Pro that Google positions for high reasoning and strong multimodality. 10 categories is how we will judge it. Each category will be scored 0 to 10 and the total possible is 100. Here's what a 10 means in each category.

0

79.638 - 103.613

They're not perfect, but they're best in class for most users. Writing is that it sounds natural, adapts tone, stays organized, and follows constraints. Coding is that it gets the correct code, good debugging, handles edge cases, explains changes, and does not invent fake libraries.

0

104.234 - 131.507

Image generation is that it can generate and edit images reliably, follows prompts, and keeps consistency across edits and outputs useful and accurate images. Problem-solving accuracy will be graded on the math, logic, multi-step reasoning, fewer confidence mistakes, and consistent answers. And the value for money will be what you actually get for the price, plus limits and practical usefulness.

0

132.196 - 155.031

Thinking mode can deliberately slow down for harder tasks. It can track constraints, plan, and self-check. Research mode can browse and cite sources, summarize faithfully, and separate facts from guesses. Voice mode is natural speaking, low latency, stable conversations, and handles technical terms.

0

155.264 - 177.608

The user interface manages files, the project's organization, exports, integration, as well as the overall friction and convenience of the platform. And finally, speed, which is the time to first give an answer and how many turns it takes to reach a correct result. Finally, we'll do a privacy spotlight.

177.628 - 204.92

We will not do dramatic claims, but we'll focus on what controls exist than what is clearly stated in the product demonstration. CPT 5.2 Thinking is positioned by OpenAI as a flagship model for deeper work, strong vision, and better performance on things like charts and UI understanding. That matters because a lot of real-world work is visual.

Chapter 3: How will the AI tools be evaluated in this series?

205.401 - 229.602

Dashboards, screenshots, diagrams, slides. Gemini, on the other hand, is built to live inside Google's ecosystems and developer stack. Google's current Gemini lineup emphasizes price performance options like Flash, plus higher capability models for reasoning, and it pairs closely with Google's image generation systems like Nano Banana Pro.

0

235.083 - 262.166

So the matchup is a tool that's heavily optimized around an instant workflow with strong multimodal reasoning versus a tool that's heavily optimized around Google ecosystem integration and a broader model item. All right, category one. Category 1's awards rank GPT 5.2 Thinking as a 9 and Gemini as an 8. Now why does GPT win?

0

262.647 - 287.216

Well, GPT 5.0 Thinking tends to produce very structured writing when you ask for it. It's strong at clear outlines, headings, and maintaining a consistent format across long responses, especially for school or professional-style outputs like reports, slide scripts, and step-by-step explanations. OpenAI explicitly positions 5.2 thinking as a more polished for complex work outputs.

0

288.137 - 304.353

However, Gemini is very close. Gemini is also strong in natural language and can write clearly. But in many workflows, the difference comes down to consistency under constraints. When you demand a very specific structure, Gemini more reliably stays locked into the format without drifting.

0

305.074 - 317.107

Gemini still scores high, it just lands slightly behind on constraint obedience in long-form writing for most users. Category 2 is Coding.

Chapter 4: What are the specific categories for judging ChatGPT and Gemini?

318.128 - 345.364

We rated GPT-5.2 Thinking out of 9 and Gemini 2.5 Pro out of 8. GPT wins because thinking mode is a real advantage for debugging and multi-file reasoning, as it encourages deliberate checking and planning. OpenAI also positions GPT-5.2 as improved for agentic tasks and deep work, and many people's coding experience improves when the model can reason carefully instead of rushing.

0

Chapter 5: How does ChatGPT 5.2 compare to Google Gemini in writing?

346.323 - 364.301

Gemini is still strong, but is slightly lower because it's very capable, but purely on coding reliability, GPT tends to edge out and tricky debugging where the model must keep multiple constraints straight, like do not change anything else, or fix only this bug, or refactor while preserving behavior.

0

364.922 - 385.799

Gemini is still top tier, but it takes a small hit for occasional drift and long constraint heavy coding prompts. Category three, image generation. GPT 5.2 thinking will rank a 9, and so will Gemini 2.5 Pro. Now, both of these models are a tie, but for different reasons.

0

386.621 - 405.521

OpenAI has a dedicated, improved ChatGPT experience focused on instruction following and precise edits, including consistency across edits. That makes it strong not only for generating images, but also for iterative changes, like change this or keep that. Gemini also scores a 9.

0

405.961 - 424.474

Google has multiple strong image paths including Nanobanana Pro through the Gemini API and Gemini image models on Vortex that support image generation and in some cases editing. The integration with Google's tooling is a big advantage for creators and developers who want repeatable pipelines.

0

425.585 - 458.191

Now neither is a 10, as both still have safety limits, occasional prompt interpretation misses, and some inconsistency on fine details. A 9 is excellent, but not perfect. Our fourth category is Problem Solving Accuracy. GPT 5.2 Thinking runs a 9, while Gemini 2.5 Pro runs a 8. GPT 5.2 Thinking is explicitly positioned as stronger on complex work and vision-based reasoning.

458.792 - 480.263

With open AI claiming substantial error rate reductions in chart reasoning and software interface understanding, that same careful reasoning orientation tends to show up in logic and multi-step tasks too, especially when the old problem has traps. Dimini scores an 8 because even though it has strong reasoning models, across typical everyday problem sets,

480.935 - 500.31

GPT tends to produce fewer confident but slightly off final answers. Gemini basically loses a point because of its average consistency in harder multi-step tasks. 10.5 is the value for money. We ranked GPT as a 7 and Gemini as an 8.

500.797 - 520.948

Gemini wins this matchup because Google's model lineup is designed around multiple price performance points, including the flash tier concept, which often translate to good capability at lower cost for many users and developers. That breadth usually improves value because you can pick the cheapest model that still solves your main tasks.

521.401 - 549.906

Now, GPT is still solid, as GPT-2 thinking can still be a strong value if your workflow depends on its strengths, which is polished long outputs, vision reasoning, and integrated creation tools. But if a user just wants fast, everyday help, the premium thinking model can be less cost-efficient compared to lighter teams. All right, the next category is thinking modes.

Comments

There are no comments yet.

Please log in to write the first comment.