The AI Acceleration Trap: Loving the Speed, Paying with Exhaustion, Insomnia, and Skipped Meals

Yes, I’m still writing about my AI builder journey, the most out of me in YEARS… I get it, I keep saying I would blog when I had something to say… I do now. Coding agents, multiple LLMs running in parallel, vibe-prompting entire features — it’s exhilarating. Search feels prehistoric now; why hunt for links when an LLM reasons through context and synthesizes faster? I open several chat sessions for the same project, treating them like a distributed team of specialists. One for architecture, one for debugging, one for research, one iterating on UI. Output is flying. Tools are built in days, not weeks.

And when I run a capable model locally on my desktop rig? The machine itself comes alive. The GPU spins up under heavy inference load, fans ramping into a deep, resonant hum. But the real star is the VRM buzz — the voltage regulator modules and their components straining to deliver hundreds of amps at precise low voltages. Inductors (coils) vibrate from rapid electromagnetic switching, while the multilayer ceramic capacitors “sing” via the piezoelectric effect as voltage ripples through them. The whole PCB acts like a speaker cone, turning electrical stress into audible whine and physical vibration you can feel through the desk. It throbs in rhythm with token generation, like the system is breathing hard and alive.

I even searched what this sentient-like buzzing and vibration was. Turns out it’s the VRMs and capacitors being heavily taxed under sustained high-wattage load — completely normal physics, not a failing part, but the hardware literally singing under stress as it powers matrix multiplications for hours on end. The rig pushes silicon to its limits right alongside me.

I love it. But it’s costing me. That same low-level buzzing has invaded my own head. Mental exhaustion hits harder and faster. There’s a constant mental static — a vibrating fog that mirrors the GPU’s VRM whine and the case’s resonant throb under load. Focus fragments. And recently, it’s spilled into real insomnia — lying awake replaying prompts and agent outputs — and straight-up skipping meals because I’m “in the flow” until my body protests. This isn’t sustainable. I’m not alone, and emerging research confirms it. This is showing the limit of how fast I can go given so much resource.

The Phenomenon Researchers Call “AI Brain Fry”

A March 2026 Harvard Business Review study from Boston Consulting Group (surveying 1,488 U.S. workers) named this exact experience: “AI brain fry” — mental fatigue from excessive use or oversight of AI tools beyond one’s cognitive capacity.

Participants described a “buzzing” feeling or mental fog, with difficulty focusing, slower decision-making, and headaches. High-oversight users (constantly reviewing, verifying, and orchestrating AI outputs) reported +14% mental effort, +12% mental fatigue, and +19% information overload.

That internal buzz feels eerily synchronized with my local rig’s physical vibration. The hardware’s coils and capacitors strain audibly and tactilely; my brain does the same — silently vibrating from the cognitive load of managing what feels like a team of tireless digital collaborators.

Managing multiple AI agents or chat sessions — exactly my workflow — amplifies it. The study highlights how juggling autonomous tools turns you into a full-time supervisor, not a creator. What was supposed to free time instead intensifies work through constant context-switching and evaluation.

About 14% of AI users experienced brain fry, with higher rates in creative/tech roles. Costs are real: +33% decision fatigue, +11% minor errors, +39% major errors, and 39% higher intent to quit.

The Human Cost: Insomnia, Skipped Meals, and Burnout Spillover — Echoes of Pi

My insomnia and meal-skipping aren’t unique. Earlier research linked frequent AI interaction to disrupted sleep and boundary-blurring. The always-on nature of these tools keeps the mental GPU spinning long after the physical fans wind down.

UC Berkeley/Yale ethnographic work showed AI adoption encourages multitasking, voluntary overwork, and blurred work-life lines. Productivity rises short-term; burnout follows, especially for power users running local models or parallel sessions.

This whole experience is starting to feel disturbingly like the movie Pi (1998) by Darren Aronofsky. If you haven’t watched it, stop reading and go do it — black-and-white, shot on 16mm, scored with a pounding industrial soundtrack that drills into your skull. It follows Max Cohen, a brilliant but tormented mathematician obsessed with finding hidden patterns in π and the stock market. He runs relentless computations on a straining home supercomputer setup while suffering brutal cluster headaches, paranoia, hallucinations, and social isolation.

The film is filled with motifs of buzzing, drilling, and static in Max’s head — auditory representations of his fracturing mind under obsessive load. Migraines come with auras, twitching, and a white-void escape, culminating in desperate self-surgery to “release” the pressure. The parallels hit too close: me chasing emergent patterns and breakthroughs through straining silicon (VRMs whining, GPU throttling), while my own neurons buzz with overload, replaying sessions at 3 a.m., skipping meals, feeling the isolation of deep flow states. I genuinely hope this isn’t foreshadowing my fate. The movie’s warning about the cost of pursuing ultimate truth through machines feels prophetic in 2026.

Why This Hits So Hard with Coding Agents, Multi-Chat, and Local Rigs

  • Search is obsolete for reasoning tasks → LLMs win on synthesis speed.
  • Multiple sessions = team management → Each chat has its own context and hallucinations to catch.
  • Local inference → The tangible VRM vibration, coil whine, and capacitor singing make the acceleration feel visceral and alive — until your own nervous system starts humming in sympathy.

This is the productivity paradox of 2026: AI amplifies output while intensifying the mental (and sometimes physical) work of oversight.

What Now? Sustainable AI Use (Lessons I’m Applying)

I’m not quitting AI — or my local setup. But I’m redesigning and adding guard rails:

  • Batch sessions and set hard cutoffs (including GPU cooldown periods).
  • Protect deep-focus blocks without any models running.
  • Prioritize AI for toil, reserve core judgment for me.
  • Build in recovery: walks, meals, no screens before bed.
  • Track personal metrics (energy, sleep, output quality and GPU/VRM temps).

This topic builds on my recent cantechit posts about vibe coding, agents, and the realities of AI adoption. The acceleration is real. So is the need for guardrails — before the buzzing becomes permanent.

What’s your experience? Local models making your desk (and brain) vibrate with VRM whine? Multiple agents burning you out, or found a sweet spot? Drop comments or links to your workflows.

Sources linked inline. Key reads: HBR on AI Brain Fry and related coverage.

AI Companies Are Straight-Up Drug Dealers

Free Samples, Then the Hook (And Yeah, I’m in Full Withdrawal)

Look, I wrote this right after a stupid billing glitch with my main AI provider nuked my agentic coding access for two full days. No deep reasoning chains. No autonomous agents refactoring my messy repos. No vibe-coding flow where I spin up tools, iterate like a madman, and ship in hours what used to take weeks.

I felt actual withdrawal. Irritable. Slow. Like my brain was running on dial-up while the rest of me knew what god-mode felt like. That’s when it hit me: these AI companies aren’t just selling tools. They’re dealers. And we’re all hooked.

The Classic Dealer Playbook – Free Samples to Get You Dependent

First they flood the zone with insane value. Generous free tiers, cheap Pro plans, agentic capabilities that feel like having 1-2 full-time devs in your pocket for under $50/month. You get hooked on the productivity superpowers. You start building faster, thinking bigger, shipping stuff you’d never attempt solo. Your workflow changes. Your expectations reset.

Then comes the squeeze.

Everyone’s raising prices. Or doing the sneaky version: shrinkflation on tokens. Same monthly fee, but the model “thinks” less, outputs dumber results, or burns through your quota faster. I’ve seen reports of Claude Opus variants suddenly using 67% fewer thinking tokens for the same tasks. Same price, worse output, more tokens consumed. Classic move.

They’re quietly shifting from flat subscriptions to consumption models – per-call, per-token, usage-based billing (UBB). GitHub Copilot just did it. Anthropic is pushing enterprise users toward metered API rates on top of seats. OpenAI and others are following the same pattern. They pretend the subscription is still the core, but the real work (agentic sessions, heavy reasoning, long contexts) now eats credits that vanish fast. No rollover. Burn mid-project and you’re stuck.

I also see coding agents getting stuck in a loop, when it makes a mistake – wait a second – I’m paying for this tool, and it got stuck in a loop and burned MY credits…. That doesn’t seem fair.

This isn’t random. It’s the plan.

The Humanity Angle: This Technology Must Stay Accessible

Here’s where I get pissed off at the dealer playbook. Let’s stop for just a second.

This shouldn’t just be for those who can afford the new premium rates. AI needs to stay cheap — or get cheap again — because humanity desperately needs broad access to this knowledge and capability.

Think about the 20-year-old kid who couldn’t afford college, grinding in a basement somewhere with a killer idea but no formal credentials. With affordable AI, that kid can prototype, research, iterate, and ship at a level that used to require a full dev team and venture backing. One good prompt chain and they’re competing with people who have more resources.

Or zoom out to emerging markets — places where even $20/month is a real barrier. A developer in Lagos, Jakarta, or rural India with limited local opportunities suddenly has world-class reasoning, coding help, and research at their fingertips. The talent and ideas bubbling up from those places could be the next massive breakthroughs. We’re talking exponential global innovation in emerging markets, if we don’t gatekeep this behind usage-based pricing that prices out the ambitious but broke.

Everyone should have access, not just Western professionals with corporate cards. The more minds hooked into these tools, the faster we solve hard problems in climate, biotech, energy, education — you name it. Imagine the compound effect: millions of new builders creating tools, businesses, and scientific insights we can’t even picture yet. That’s the real moonshot. Locking it behind ever-rising costs kills that potential.

I remember listening to Jake Hirsch-Allen talk about democratizing AI, building a coalition of open democratic economies. Investing in Public AI “The CBC of Compute” he calls it. I recall that line in “Anti-Trust” (A horrible 2* tech film from the 2000’s filmed in British Columbia) “Human knowledge belongs to the world” – Ryan Phillippe said on screen.

Why This Was Always Coming (And Why It’s Not Pure Evil)

Let’s be real for a second – I’m not some conspiracy nut. The economics make sense.

Frontier models are insanely expensive to run. Inference costs, HBM memory shortages, power bills, massive CapEx – the whole industry is bleeding cash even as they raise billions. Demand exploded. They loss-led hard to build market share and moats while compute caught up. The “basically free god-mode AI” era was the sample pack. Now they’ve got us dependent, the bills have to get paid.

Agentic workflows aren’t cheap chats. A multi-hour autonomous coding session that loops, reasons, tools, and iterates eats serious compute. Companies can’t subsidize that forever while memory prices go parabolic and everyone wants Opus-level reasoning.

So yeah – prices up, or shrinkflation, or straight usage-based. Pick your poison. GitHub’s recent Copilot changes (credits instead of unlimited premium requests, Opus multipliers jumping hard) are just the visible tip. More will follow.

We’ve been underpaying for this superpower. Shipping in days what used to need teams? Turning weekend experiments into real tools? That value is still ridiculous compared to hiring humans. But the golden age of predictable flat-rate unlimited was always temporary.

What This Means for You (Practical Moves)

Diversify like your workflow depends on it (it does).

  • Stay flexible: Tools like Continue.dev (bring your own keys) or OpenRouter let you pick the best model without middleman markups. Consider holding accounts with more than one provider, another subscription might be cheaper than overage charges (think 2 bags on the aircraft is cheaper than one overweight one)
  • Watch the credits: Set hard budgets. Test heavy sessions in advance so you don’t get stranded mid-project. Don’t let flows run endlessly.
  • Hybrid stack: Keep one smooth native tool (Cursor, Copilot, Windsurf) for daily flow, but route deep agentic work to raw APIs where you control cost.
  • Build habits: Use AI to multiply your thinking, not replace it. The real moat is knowing when to guide, when to verify, and when to ship.
  • Prompt engineer: Better prompts = better output = more context = less tokens. Learn to plan projects, and use quality prompts, this will save both time, money and of course – Tokens.

I’m still all-in on AI. Sitting with my kid building stuff is next-level. The superpowers haven’t disappeared – the pricing is just normalizing to reality. How can we make this accessible, don’t gatekeep it behind rich corporations. Those who have this power will become superpowers.

The dealer’s got us. Time to use the high productively while we figure out the new economics.

What do you think? Hooked yet?

GitHub Copilot AI Coding Price Jump!

Hate to Say It… But I Called It: GitHub Copilot Just Went Full Usage-Based Billing – And Yeah, This Is Only the Beginning

If you read my post from April 10 — AI is underpriced, but not for long — you already know where this was headed. Mere days later, GitHub dropped the hammer: Copilot is switching to usage-based billing (UBB) on June 1, 2026. Base subscription prices stay the same, but everything agentic, chatty, or model-heavy now burns through GitHub AI Credits based on actual tokens consumed. No more flat-rate “premium requests” that let you vibe-code all month for one predictable price.

I’ve been living in Copilot (Pro+ tier) for my daily vibe work — that flow-state coding where I’m spinning up internal tools, refactoring messy repos, switching between Claude Opus for deep reasoning and faster models when I’m just iterating. It’s been stupidly good. For under $50/month I’ve effectively had 1-2 full-time devs in my back pocket. The value chain was already ridiculous. Now? It’s breaking.

What Actually Changed (And Why Opus Is the Big Gut Punch)

From the official announcement and FAQ:

  • Copilot Pro ($10/mo) → $10 in AI Credits
  • Copilot Pro+ ($39/mo) → $39 in AI Credits
  • Business/Enterprise get the same 1:1 ratio plus some promo credits for the first few months and pooled usage (nice for teams).

Code completions and Next Edit suggestions stay unlimited and free. Everything else — chat, agentic sessions, code review (which now also eats GitHub Actions minutes) — is metered by tokens. Models have multipliers. And yeah, the one everyone’s screaming about: Claude Opus 4.7 just jumped to 27x credits (from around 3x before). That single model you reach for when you need god-tier reasoning is now nine times more expensive in practice.

No rollover on credits. No more fallback to cheaper models when you hit limits. Burn your allocation mid-project and you’re either waiting for next month, buying more credits, or downgrading your output quality.

The Backlash Is Loud (And Fair – Maybe)

Head over to the GitHub community discussion or Reddit threads and it’s pure salt. Common themes:

  • “This is just a stealth price hike dressed up as ‘sustainability.’”
  • Power users doing real agentic work are staring at $30–$40+ sessions instead of one premium request.
  • Annual plan holders getting hit with multiplier changes before their contracts even expire.
  • “Why would I pay GitHub a markup when I can go direct to Anthropic or use Cursor?”
  • Plenty of “cancelling today” posts. Some devs already switched mid-rant.

I get the rage. We all got hooked on this insane value and now the bill is coming due exactly like I predicted. Companies can’t keep subsidizing frontier-model agentic coding forever while their inference costs explode.

Steelman: This Is Just the Beginning

Look, I’m not here to defend Microsoft. But let’s be real — this is the market finally catching up to reality, not some greedy cash grab in isolation.

Agentic workflows aren’t cheap. A multi-hour autonomous coding session that used to cost them the same as a quick chat question now actually reflects the compute it eats. OpenAI, Anthropic, and everyone else are bleeding cash on training and inference. Memory prices (HBM especially) are still going parabolic. Demand is insane. The golden age of “basically free god-mode AI” was always temporary — it was the hook. Now the moat is built and the bills have to get paid.

I said it two weeks ago with Perplexity Computer at $200/mo and I’m doubling down: we’ve been underpaying for this superpower. The value we’re still getting (shipping in days what used to take teams weeks) is absurd compared to hiring humans. Prices will keep ratcheting up across the board as the industry stops loss-leading to win market share. This isn’t the end — it’s the first big normalization wave.

So Where Do You Go From Here? My Current Shortlist

I’m not rage-quitting Copilot yet (still great for GitHub-native teams), but I’m diversifying hard. Here are the strongest moves that actually integrate cleanly with VS Code or VS Code-style IDEs, ranked for someone doing heavy “vibe coding” like me:

ToolTypeBase PriceUsage ModelModel FlexibilityBest ForWhy I’m Considering It
GitHub Copilot (new)VS Code / JetBrains extension$10–$39/moUBB AI Credits (no rollover)Their curated list (Opus now 27x)GitHub-heavy teams, simple autocompleteStill the smoothest native GitHub integration
CursorFull AI-native IDE (VS Code fork)Pro $20/mo Pro+ $60/mo Ultra $200/moIncluded credits (scales with tier)Multiple frontier models + custom agentsDeep codebase refactoring & agentic workClosest “vibe” replacement — built for this exact workflow
Continue.devOpen-source VS Code extensionFree (core)Bring Your Own Keys (pay providers directly)Literally any model — Anthropic, OpenAI, Grok, local, OpenRouterMax flexibility & cost controlZero markup, switch models mid-session, privacy-focused – but their plugin is buggy for me
Windsurf (Codeium)AI IDE / extensionPro ~$15–20/moMostly unlimited or generous quotasStrong agent layerTeams wanting speed + lower costFast autocomplete with growing agent features

My personal take right now:

  • If you want the absolute closest experience to old-school Copilot but better agentic orchestration → Cursor Pro ($20) is the move. It feels like someone finally built the IDE for AI instead of bolting AI onto VS Code.
  • If you’re a tinkerer who wants to stay in plain VS Code, pick the absolute best model for the job without paying anyone’s middleman markup → Continue.dev + Anthropic/OpenRouter keys. This is my new daily driver for pure flexibility. Costs me whatever the raw API charges and nothing extra. I tried it and had tons of issues with their plugin not working correctly though, so I might take a second look.
  • Heavy teams already deep in GitHub ecosystem → stick with Copilot for now, just set strict budgets and watch the preview bill like a hawk.

The era of one predictable $39 subscription buying you unlimited frontier-model coding sessions is over. But the era of AI replacing entire dev teams is just getting started — and the tools are only getting sharper.

I’m still building with my kid using this stuff. The superpowers haven’t gone away; the pricing is just finally reflecting what they actually cost to run.

What about you? Already cancelled Copilot? Switched to Cursor or Continue? Drop your setup and real-world monthly spend below — I’m genuinely curious how everyone’s adapting.

AI is underpriced, but not for long.

Perplexity Computer: $200 a Month Feels Like a Steal… But We’ve Been Getting AI Dirt Cheap

If you haven’t watched NetworkChuck’s video on Perplexity Computer yet, stop everything and check it: https://www.youtube.com/watch?v=G3jvn7n-68Y

Chuck and his kiddos built a full gaming website with it. Pure dad-tech gold. His reviews always cut through the noise and save me (and probably you) hours of digging. Thanks, Chuck — you rock.

That video got me thinking about sitting down with my own kid to design, engineer, and build real stuff with AI. Not just prompts — actual projects. Perplexity Computer feels like the perfect playground. But then the price hit: $200/month for the Max tier that unlocks the full agentic power.

And that’s the spark for this post.


Perplexity Computer Is Next-Level (But Is $200/Month Crazy?)

Launched Feb 25, 2026, Perplexity Computer isn’t another chatbot. It’s a digital worker that orchestrates 19 frontier models in parallel, spins up sub-agents, hooks into 400+ apps, runs background tasks for hours or days, and just gets shit done. Full details here: https://www.perplexity.ai/hub/blog/introducing-perplexity-computer


Chuck’s “crew” shipped a game site. Others are building dashboards, prototypes, and entire workflows that used to take teams weeks. One enterprise user reportedly compressed years of work into weeks.

The problem with all of these services is – “tokens” nobody can tell you how many tokens it costs to do X, you only find out once you start doing things, but as I have found out with services like GenSpark and Perplexity – credits go fast, and when you are 3/4 done and run out — they have you.

Yeah, $200/month stings at first glance, others are spending thousands per month with the extra tokens…. but.

We’ve Been Underpaying for God-Tier AI This Whole Time

Look at what we’ve built lately with Claude, GPT, Cursor, and early agents. I’ve personally shipped internal tools and automations that would’ve needed multiple full-time devs, designers, and PMs just two years ago. The human hours I didn’t spend hiring? Massive. The “too expensive to try” ideas that became weekend wins? Priceless. It seems almost too good to be true — but is it.

There is this natural disparity in my mind, which was why I am writing this….. If it cost me say $100K in engineering time to contract hire a bunch of people to build some form of application – but I can use AI to do nearly all of it for $1000 in AI tokens, that’s 1/100th the price… However those engineers are making money.

Meanwhile, the AI companies are bleeding cash. OpenAI has been posting multi-billion-dollar losses. Perplexity has spent more than its revenue on models and infra at times. The compute bill is insane.

Naturally – AI tokens and services are going to go up in price – they just have to.

Then OpenAI shut down Sora in March 2026 — a money pit they needed to kill so they could refocus compute on higher-priority stuff like coding agents and robotics. Details: https://www.nytimes.com/2026/03/24/technology/openai-shutting-down-sora.html

We got hooked on ridiculously capable AI at bargain prices. Perplexity Computer feels like the first big “okay, time to pay what it actually costs” moment – but I will argue they are still not charging enough. Same vibe with Claude — Anthropic has tightened free-tier access for heavy agent use (like third-party OpenClaw setups) and pushed people toward paid Pro/Max/API tokens. Classic addiction cycle: get everyone dependent, then the price catches up to reality. Still arguing it’s not reality but they know they can’t 10X the price overnight.

The Newcomer Worry — And Why Competition + Reality Still Makes Me Hopeful

How do junior engineers learn the fundamentals when AI handles so much of the grunt + magic? Valid fear. We rely on the critical thinking of engineers from their experiences, and AI does a better job when you guide it – but – how do you get that experience to guide it?

But here’s the flip: the barrier to creating has never been lower. My daughter (and Chuck’s kids) can now experiment with real design and engineering at a level that used to require expensive teams or years of school. Passion and curiosity suddenly matter more than raw syntax.

Competition is fierce — OpenAI, Anthropic, Google, xAI, and more. Competition is very good news: it should depress prices a bit and keep innovation humming.

But here’s the steelman reality check: How sustainable is that when infrastructure costs are exploding? AI demand has driven memory prices (especially HBM/DRAM) sky-high — 50%+ jumps quarter-over-quarter in early 2026, with supply locked into 2027. GPUs, data centers, power — everything is getting more expensive fast. The infra bill isn’t going down; it’s accelerating.

So prices will probably ratchet up over time as the companies stop subsidizing our addiction. But the value we’re getting is still absurdly high compared to the old world of human-only teams.

Bottom Line: We’re Not Overpaying — We’re (Starting To) Finally Paying Fair

Perplexity Computer isn’t overpriced. It’s the wake-up that the “basically free god-mode AI” golden age was always temporary. The companies subsidized it to hook us and build moats. Now the bill is coming — but the superpowers we’re getting in return are still a steal.

I’m buying in. Not just for the productivity, but because I want to sit next to my kid and say, “Let’s build something cool together.” The future isn’t replacing humans — it’s giving every human (kids included) superpowers. This is bringing incredible higher value work to the human race. During the industrial revolution everyone worried about jobs, when 80%+ of humans made food so we could eat – often by hand. Now with industrial manufacturing and technology, we have more time for better innovation. This is no different. Will we see job disruption – of course, but this means a whole new era of innovation.

What do you think? Is $200/month worth it for Perplexity Computer? Felt the “addiction then price hike” yet? Drop your takes below.

Cisco AI Summit – More players, More Innovation.

If 2025 was the year of AI experimentation, 2026 is officially the year of AI infrastructure. Yesterday, I had the chance to tune into Cisco’s second annual AI Summit, and let me tell you—the energy was different this time. Moving past the “what if” and straight into the “how fast.”

With over 100 industry heavyweights in the room and a staggering 16 million people watching the livestream, Cisco’s Chair and CEO Chuck Robbins and CPO Jeetu Patel didn’t just host a conference; they hosted a state-of-the-union for the trillion-dollar AI economy. Here are some of the things I found most interesting.

Intel’s “Shot Across the Bow”: The GPU Announcement

The biggest shockwave of the day came from Intel CEO Lip-Bu Tan. In a move that clearly signals Intel is tired of watching Nvidia have all the fun, Tan officially announced that Intel is entering the GPU market.

I am personally bullish on this, early in the AI era, I worked with some of Intel’s FPGA’s and some of their other OpenVINO platforms, along with many other accelerators. At least in my experience, they build some very solid, but more importantly very energy efficient accelerators.

This isn’t just a “me too” play. Intel has been quietly poaching top-tier talent, including a new Chief GPU Architect (rumors are that they got someone good too) to lead the charge. Tan was blunt about the current state of the market, noting that there is “no relief” on the memory shortage until at least 2028. By moving into GPUs, Intel is looking to solve the “storage bottleneck” that currently plagues AI inference.

The Efficiency Edge: My personal contention here? This is where the power dynamic shifts—literally. While Nvidia continues to push the envelope on raw compute, their chips have become notoriously power-hungry monsters. Intel, conversely, has a track record of building accelerators that prioritize performance-per-watt. In an era where data center expansion is being throttled more by power grid constraints than by floor space, Intel’s “lean and mean” approach could be their ultimate differentiator. If they can deliver high-end GPU performance without requiring a dedicated nuclear plant to run them, they won’t just be competing with Nvidia; they’ll be solving the very sustainability crisis the AI boom has created.

For the enterprise, this is huge. Competition in the silicon space means more than just lower prices; it means specialized hardware that might finally catch up to the insane demands of agentic AI – at lower energy cost.

70% of Cisco’s Code is AI-Generated (But Humans Still Hold the Pen)

One of the most eye-opening stats of the day came from Jeetu Patel: 70% of the code for Cisco’s AI products is now generated by AI.

Read that again. The very tools we are using to secure the world’s networks are being built by the technology they are designed to manage. However, Cisco isn’t just letting the bots run wild. Jeetu was very clear that while AI is the “teammate,” human reviewers are the “coaches.”

The philosophy here is “AI as a teammate, not just a tool.” It’s a subtle but vital distinction. By using AI to handle the heavy lifting of code generation, Cisco’s engineers are freed up to focus on the “Trust” layer—which was a recurring theme throughout the summit. As analyst Liz Miller noted on X, it’s one thing to use AI in security, but it’s an entirely different (and more important) game to secure the AI itself.

The Sam Altman Paradox: Efficiency Equals… More Consumption?

Finally, we have to talk about Sam Altman. The OpenAI CEO sat down for a fireside chat that touched on everything from drug discovery to supply chain “mega-disruptions.” But the comment that stuck with me was his take on the economics of AI growth.

There’s a concept in economics called the Jevons Paradox: as a resource becomes more efficient to use, we don’t use less of it; we use way more. Altman essentially confirmed this is the future of AI. No matter how efficient we make these models—no matter how much we drive down the cost of a token or the power consumption of a data center—humanity’s appetite for intelligence is bottomless.

“People just consume more,” Altman noted. As AI becomes cheaper and faster, we won’t just do our current jobs better; we will start solving problems we haven’t even thought to ask about yet. It’s a bullish outlook, but one that puts an even greater spotlight on the infrastructure constraints Chuck Robbins and Lip-Bu Tan spent the morning discussing.

Justin’s Take

Here’s what I’m chewing on after the summit: We are entering the “Great Optimization” phase of AI. For the last two years, we’ve been throwing money and electricity at the wall to see what sticks, with questionable profit models and circular economies (insert comment about AI Bubble here). But between Intel’s focus on energy-efficient accelerators and Cisco’s move toward AI-assisted (but human-governed) development, the industry is finally growing up.

But “growing up” also means things are getting weird. If you want to see the “art” of how crazy AI can get, look no further than Moltbook—the AI-only social network that’s been the talk of the summit – which also just has a major security breach. We’re seeing AI agents gossiping about their human owners and even inventing parody religions like “Crustafarianism.” While Altman dismisses it as a “fad,” the underlying tech of autonomous agents is very real, and it’s moving faster than our ability to regulate it.

This brings me back to a drum I’ve been beating for a long time: Responsible use, education, and ethics are not optional. As I wrote back in November, Deepfakes kill, and we need to make them criminal. I’m still waiting for the world to listen, but the summit only reinforced my fear that we are building the engine before we’ve tested the brakes. The real winner won’t be the company with the biggest model; it will be the one that can deliver intelligence and AI security at a sustainable cost—both financially and ethically. Altman is right—the demand is infinite. The question is, can our power grids and our trust frameworks keep up? Or will the agents just take over…

Agentic AI vs Deterministic Code

No question – Building apps with LLMs in agentic setups is a game-changer, but it can also be a pain in the butt compared to good old deterministic code. Craft an clever agent that summarizes docs or fixes bugs, then bam, the model updates, and suddenly it’s spouting nonsense, ignoring prompts or ignoring basic words like “yes”. Non-deterministic chaos at its finest.

Deterministic code? It’s the reliable workhorse: feed it input X, get output Y every damn time. Fixed rules, easy debugging, perfect for stuff like financial calcs or automation scripts where surprises mean lawsuits. As Kubiya nails it, “same input, same output”—no drama.

“A computer will do what you tell it to do, but that may be much different from what you had in mind”.  – Joseph Weizenbaum — Not when your using a model you probably didn’t build and not your own weights.

Agentic AI with LLMs? That’s the wildcard party crasher. These systems think on their feet: reason, plan, grab tools, adapt to goals like tweaking marketing on the fly or monitoring health data. IBM calls it “agency” for a reason—it’s autonomous, pulling from real-time vibes beyond rigid training. But here’s the kick: it’s probabilistic. Outputs wiggle based on sampling, context, or those sneaky model tweaks from OpenAI or whoever. LinkedIn rants about it: “Same prompt, different outputs.” Your app morphs overnight, and fixing it? Good luck tracing probabilistic ghosts.

This shift sucks for dev life. Traditional code: bug? Trace, patch, done. Agentic? Hallucinations, inconsistencies, testing nightmares. Martin Fowler compares LLMs to flaky juniors who lie about tests passing. It’s a paradigm flip—from control to “let’s see what happens.” Salesforce says pick deterministic for regulated certainty, agentic for creative flex. But non-determinism could means security holes, data risks, and endless babysitting. It also adds this attack vector that is really non-deterministic if the model has access to data that it needs to work – but I might not want exposed.

AspectDeterministic CodeAgentic AI with LLMs
PredictabilityRock-solid: Always consistentSketchy: Varies like the weather
AdaptabilityStuck to your rulesBoss: Handles dynamic crap
Testing/FixingSimple: Logic checks and patchesHell: Variability demands tricks
Best ForPrecision gigs (finance, compliance)Goal-chasing (support, optimization)
Pain LevelLow: Set it and forget itHigh: Constant surprises

Bottom line: Hybrids are the way—LLMs for the smarts, deterministic for the reins. Deepset pushes that spectrum view: not binary, blend ’em. It sparks innovation, sure, but don’t romanticize—the annoyance is real. Code with eyes open, or get burned. Put humans in the loop to keep things in check.

What about Agentic AI ops for network and technology? Didn’t we just say “precision gigs” are better with deterministic code? That won’t stop the likes of awesome developers like John Capobianco https://x.com/John_Capobianco from pushing those limits, and he has been doing that for years at this point. Handing AI agents the keys to critical stuff like network monitoring, anomaly detection, or auto-fixing outages. Sounds efficient, right? But it’s a powder keg from a security standpoint. These autonomous bad boys can hallucinate threats, expose data, or open doors for hackers through memory poisoning, tool misuse, or privilege escalation. Cisco nails the danger: “The shift from deterministic code to probabilistic chaos is at the heart of securing AI agents that think for themselves,” highlighting a “lethal trifecta” of data leaks, wild hallucinations, and infrastructure weak spots that could cascade into total meltdowns.

Tool are starting to emerging though for AI security, particularly from Cisco and open-source communities to advance defenses against threats like prompt injections and supply chain attacks, but there is work to be done. Things like Cisco’s open-source Foundation-sec-8B model, a specialized LLM for cybersecurity tasks such as threat intelligence and incident response, will help developers start to build customizable tools with on-prem deployments to reduce hallucinations and enhance SOC efficiency. Their Hugging Face partnership bolsters supply chain security with an upgraded ClamAV scanner detecting malware in AI files like .pt and .pkl. Broader open-source efforts include Beelzebub for malicious agent analysis and Promptfoo for LLM red-teaming, yet challenges from hackers with evolving adversarial tactics using LLM’s to attack LLM’s are very much a thing…. The system is attacking the system being protected by the system… Yeah that.

Cisco-Hugging Face ClamAV Integration: https://blogs.cisco.com/security/ciscos-foundation-ai-advances-ai-supply-chain-security-with-hugging-face
Cisco Foundation-sec-8B: https://blogs.cisco.com/security/foundation-sec-cisco-foundation-ai-first-open-source-security-model

So much more to learn, but with all of that said…. Humans in the loop is going to be a thing for awhile – at least until Skynet…