
Last Thanksgiving, I spent an entire weekend building an app in Lovable. The math wasn't accurate. I couldn't rely on it. Ended up hiring a developer to finish.
This Thanksgiving, I built something better in a couple of hours. Embedded AI. Deployed to the cloud. Shareable. No developer.
Ethan Mollick captured this shift perfectly in his Substack, One Useful Thing: "Three years ago, we were impressed that a machine could write a poem about otters. Less than 1,000 days later, I am debating statistical methodology with an agent that built its own research environment."
That's not incremental improvement. It's a phase change. And we crossed it this month.
I've felt what Mollick calls the "Jagged Frontier" – where AI is brilliant at some tasks and frustratingly unreliable at others. He recently wrote that "the era of the chatbot is turning into the era of the digital coworker."
His research with BCG showed that below-average performers improved by 43% when using AI. A skill-leveling effect that disproportionately helps non-experts reach expert-level output.
We've crossed that threshold. The frontier is still jagged, but the floor has risen dramatically.
The Significance of What Was Released
In the last 60 days, the releases have been relentless: Claude Opus 4.5, Gemini 3, GPT-5.2, Lovable 2.0 with Agent Mode, Claude for Excel, NotebookLM upgrades, native file creation with context compaction. Any one of these would be notable. Together, they represent something bigger.
Claude Opus 4.5 deserves a closer look. On ARC-AGI-2, a test of abstract reasoning that requires solving novel problems from just a few examples, Opus 4.5 scored 37.6%. That's more than double GPT-5.1's 17.6%. On multi-step tasks, testers noted it "doesn't lose the plot." It holds instructions and constraints across long workflows where other models start making routing mistakes. And pricing dropped by two-thirds: from $15 per million input tokens to $5.
The enterprise results are even more striking. Box ran their advanced reasoning eval on Opus 4.5 and saw a 20 percentage point improvement over Opus 4.1, which came out just three months ago. As Aaron Levie put it: "This eval gets closer to approximating what a knowledge worker does as a discrete task with their enterprise documents. It could be a financial analyst that's analyzing a company or a consultant doing research for a client."
On industry-specific tasks, the high-effort model hit 96% accuracy in education, 89% in energy, and 66% in healthcare and life sciences.
Beyond the benchmarks, Opus 4.5 feels different. It follows the instructions in my custom Claude projects more reliably. There are fewer hallucinations when I double-check. And for someone who constantly runs into the end of a thread, even with a 200K context window, the new context compaction is a game-changer. If I’m doing real strategic planning for a client deliverable, I no longer have to summarize and restart in a new chat just to keep the conversation going. The entire arc stays intact.
Gemini 3 Pro is equally impressive in different ways. It scored 91.9% on GPQA Diamond, a PhD-level scientific knowledge benchmark. Highest of any model. On AIME 2025, a tough math competition, it hit 95% without tools and 100% with code execution. Long-context performance (77%) crushed GPT-5.1 (61%) and Claude Sonnet 4.5 (47%). This matters for anyone working with large documents or complex research. It also topped the WebDev Arena leaderboard for frontend development.
Gemini 3 includes Nano Banana Pro, Google's image generation model optimized for accurate text rendering. That capability now powers NotebookLM, which added Source-to-Slides, Instant Infographics, and Video Overviews. Drop in a proposal or case study, and it generates branded visual assets in seconds. No more garbled labels or misspelled headers. The visual output barrier that used to require a designer has collapsed.
And then, on December 11th, OpenAI released GPT-5.2. I haven't had time to fully test it yet, but the benchmarks are impressive. On GDPval, which measures real-world knowledge work across 44 occupations, GPT-5.2 Thinking beats or ties human experts on 70.9% of task, according to expert human judges. These tasks include making presentations, spreadsheets, and other artifacts.
Box ran their enterprise eval and saw reasoning scores jump from 59% to 66%, with even bigger gains in specialized verticals like media (81%) and financial services (71%). OpenAI claims it produces outputs at 11x the speed and less than 1% the cost of expert professionals. The race isn't slowing down.
Lovable 2.0 shipped Agent Mode, which is now the default. It debugs autonomously, plans multi-step tasks, and fixes its own errors. Last Thanksgiving, I needed a developer to finish what I started. This Thanksgiving, the agent handled it.
Lovable is what I use, but it's not alone. The entire category has leveled up in the last 90 days. Replit now has a Design Mode powered by Gemini 3 and Nano Banana Pro. You can generate polished dashboards, landing pages, and marketing visuals directly from a prompt, with accurate text rendering that actually looks professional. Bolt optimizes for speed when you need a quick prototype. v0 generates pixel-perfect UI components from a chat prompt. All of them now offer one-click deployment.
What does this mean for a salesperson or consultant?
You can build a custom ROI calculator for a target account. A personalized landing page for an ABM campaign. An interactive demo tailored to a prospect's specific use case. Tools that would have required a developer, or at least a technical friend, are now at the tip of a well-written prompt. Twenty to thirty bucks a month.
That cost trajectory matters. The intelligence is already past the threshold needed to transform knowledge work. What's lagging is deployment: getting these capabilities into the hands of average workers through agents, integrations, and workflows. But that bottleneck is eroding fast. As costs keep falling, the same reasoning power that today requires a power user to access will be embedded in tools everyone uses.
Early adopters get the advantage now. Everyone else will get it later, when it's cheaper and more seamless. But the capability is here.
The Four Pillars of Professional Output
Here's what hit me this Thanksgiving weekend: AI is now better than the average knowledge worker at the core tools that define knowledge work.
Think about those tools. The typical knowledge worker’s productivity suite, whether in Microsoft Office or Google Workspace, comes down to documents, slides, and spreadsheets. Then there's the software we use and build. Four pillars in total. AI has gotten meaningfully better at all of them.
Documents. AI has been strong at writing for a while. But the workflow was always fragmented: brainstorm in Claude, refine in Google Docs, format in Word (if you must), fight with version control. Now I can brainstorm a client proposal, refine the messaging through multiple iterations, and export a branded .docx. All in one conversation. Rough idea to polished deliverable, one thread.
Slides. This is where the change feels most dramatic. NotebookLM's Nano Banana upgrade can turn source documents into branded slide decks with accurate text rendering. No more garbled labels or misspelled headers. Infographics that would have required a designer now generate in seconds. Tools like Genspark and Claude's native file creation mean I can go from "here's my thinking on this deal" to "here's a 12-slide deck for the executive sponsor" without leaving the conversation.
Spreadsheets. This weekend I used Claude for Excel to build a complex waterfall revenue model. It included sales cycle inputs, rep commission structures, and variations for account managers versus new business reps. I don't know Excel formula syntax. I couldn't write a VLOOKUP from scratch if you paid me. I built the model anyway. Described what I wanted in plain English, let Claude write the formulas. When I needed a variation that accounted for different commission tiers, I just asked. Three minutes later, I had it.
But it's not just building from scratch. Claude for Excel understands your entire workbook: nested formulas, multi-tab dependencies, all of it. You can ask "what assumptions drive the revenue forecast in Q3?" and get an answer with cell-level citations. You can test scenarios ("increase growth by 2% and show the impact on terminal value") without breaking anything. You can debug #REF! errors that would have taken an hour to trace. If you've ever inherited a complex model and felt lost, this changes how you work.
Software. I mentioned Lovable in the intro – my experience this Thanksgiving versus last Thanksgiving. That contrast tells the whole story. What took a weekend and a developer now takes a couple of hours and no outside help. Lovable's Agent Mode debugs autonomously, plans multi-step tasks, and fixes its own errors. The gap between "I have an idea for an app" and "here's a working prototype" is gone.
And it's not just app building. In the last month, Zapier and Lindy have both released the ability to create automations using natural language. You can describe what you want ("analyze all my sales calls each week, identify patterns, and send a summary report to managers and the marketing team for messaging insights") and have it build the full automation. That used to require hiring someone technical. Now you just describe it clearly.
What This Means for Sales
The average salesperson or knowledge worker was never 100% proficient across all these tools. Most couldn't build complex financial models with nested formulas. Most couldn't design branded decks that look like they came from marketing. Most couldn't deploy a functional ROI calculator as a web app.
Now they can. If they learn to prompt effectively.
That's the Rubicon. It's not that AI has become marginally better. It's that AI capability now exceeds the skill level of the average knowledge worker across the core tools of the job. The compound effect of all these releases in a single 60-day window pushed us past a tipping point.
Every time I finish one of these projects, I'm struck by how much has changed. I've been a power user for years, pushing these tools to their limits. If it feels like a step change to me, imagine what it means for someone just getting started.
McKay Wrigley, a well-known AI developer and content creator, captured this feeling perfectly in a recent post about Claude Opus 4.5: "After a few of these experiences your brain realizes 'oh. ok. we live in this world now.' And then you're hooked. From that moment on, you'll never work the same way again."
The Strategic Question Has Changed
It's no longer "Is AI capable of producing above-average outputs?"
It's "How fast can you learn to deploy these tools before your competitors do?"
Diffusion throughout the economy will take time. Organizations move slowly. But for individuals (salespeople, entrepreneurs, consultants, managers) the window is open right now. The tools are available, and the capability is there. The only variable is how quickly you and your team learn to use them.
Your competitor who figures this out six months before you do will operate at a different level. They'll produce proposals faster, build custom tools for prospects, generate visual assets on the fly, model deal economics in real time.
Not because they're smarter. Because they learned to prompt.
Don't Believe Me? Try This Today
Take any dense document you've been meaning to do something with: a proposal, a case study, a QBR summary, a competitive analysis, a research report. Drop it into NotebookLM. Then try one of these:
For Sellers:
"Create a one-page visual summary of this document for a VP-level buyer. Highlight the 3 most important takeaways, include one data point that matters, and make it skimmable in 30 seconds. Style: clean and professional. Ensure all text is legible."
For Sales Leaders:
"Turn this document into a briefing I can share with my team. Summarize the key insights, flag anything that requires action, and format it as a quick-reference guide. Style: clear headers, bullet points where helpful."
The output won't be perfect. Treat it as Version 0.1 – a draft to refine, not a final deliverable. But you'll have a visual asset in 30 seconds that would have taken hours (or a designer) before.
That's the Rubicon. The question is whether you're crossing it too.
If you want these capabilities wired into your sales process – playbooks, prompts, and training that actually moves revenue – reach out to Dana Consulting and we’ll map out a practical plan.
