The BS-Free Context on AI Development

Context: Skills and Intent

My background sits across software engineering, product management, infrastructure operations, licensing models, and commercial delivery. I’ve operated close enough to the system to understand how things actually move when they leave the whiteboard:

  • How delivery teams behave under pressure.
  • How infrastructure decisions quietly compound cost.
  • How licensing models shape behaviour long after procurement signs off.

I wanted to step inside it, using my own constraints, to evaluate: What actually changes when AI becomes the developer… and what doesn’t?

The Objective: From Concept to Commercialisation

The brief was simple, but intentionally uncomfortable. My goal was to:

  • Build an iOS app.
  • Use AI as the primary developer.
  • Operate from product and architecture thinking.
  • Move toward something that could actually be commercialised.

And importantly:

  • I hadn’t been deeply hands-on coding for years.

So this wasn’t about “AI makes developers faster.” This was:

  • Can someone with product, architecture, and delivery experience – without recent coding depth – use AI to build something real?

The Structure

  • Two weeks prototyping the technology on any random idea that comes to mind.
  • Then take the learnings and start fresh on a new project while leveraging as much as possible from the first build.
  • No extended planning phase. No perfect documentation upfront. Just enough structure to move.

Phase One: The Illusion Feels Real (Because It Is… at First)

The first few days feel like a cheat code. In my case, I sat inside Xcode, working with OpenAI Codex, and the feedback loop is immediate.

You prompt something like:

“Create an app that does this and I need notifications to this, simple data persistence to that…”

And it doesn’t just give you snippets. It scaffolds:

  • SwiftUI views.
  • Data models.
  • Persistence layer (Core Data / local storage).
  • Basic navigation flow.

You ask: “Why this structure?” It explains. You challenge: “Can we simplify this?” It refactors.

The Speed Is Not Incremental

This is not 10% faster. This is:

  • From zero to working UI in under an hour.
  • From idea to functional prototype in a single evening.
  • From “I haven’t coded from scratch in years” to shipping something usable.

And this is where something important clicks: The bottleneck has moved.

  • It’s no longer execution.
  • It’s decision-making.

The Real Unlock: Experience Multiplies the Output

AI doesn’t level the playing field. It stretches it.

Because I’m not asking: “How do I write this function?” I’m asking:

  • “What’s the right structure for this app?”
  • “How do I keep this local – first but extensible?”
  • “What happens when I scale this?”

AI fills in execution. But direction still comes from experience.

The Sweet Spot: Where AI Is Almost Perfect

There’s a zone where everything just works.

  • Small apps.
  • Clear scope.
  • Limited risk.
  • Minimal integration complexity.

In this space, AI is exceptional. The loop becomes: Build → Break → Fix → Understand. And learning compresses into execution.

Then Iteration Arrives (And the Tone Changes)

Week two introduces friction. Not failure. Friction.

  • Context breaks.
  • Refactors drift.
  • Behaviour becomes inconsistent.
  • You repeat yourself more than expected.

And this is where the real work begins.

From Prompting to Managing

You stop chatting. You start structuring.

  • Prompt templates.
  • Guardrails*.
  • Defined flows.
  • Context control.

AI becomes something you manage, not something you use.

BS detection tip for decision makers: 

*Guardrails – this is where the marketing noise peaks. Slide decks position it as a breakthrough control layer. In reality, it’s mostly standard architecture and safety practices any experienced engineer or architect would already put in place. It’s been repackaged, renamed, and sold back as something new.

Push for specifics. That’s where it usually unravels – vague language, recycled ideas, and very little that holds up in real-world complexity. If the person pitching it can’t clearly explain how it actually works, that’s your signal to politely ask them to leave before you replace your engineering team.

The Invisible Constraints (my use case): Xcode Is Incredible… and Limited

Inside Xcode with OpenAI Codex, everything feels tight. Not just visually. Operationally.

The loop is compressed in a way that makes the whole experience feel more capable than it might actually be in a messier environment. Apple controls the shell. The tooling is curated. The workflow is narrow enough to stay coherent. That matters more than people admit.

What you are really feeling in that moment is not just model quality.

  • You are feeling the power of an opinionated product.
  • The right abstractions are already in place.
  • Enough engineering complexity is hidden to keep you moving, but not so much that you feel blind.

It sits in that sweet spot between real engineering and low-code convenience, where the system removes friction without fully removing agency.

That polish is not neutral. Closed platforms do not just optimise for usability. They optimise for retention, partner leverage, and revenue capture across a broader ecosystem. Subscription products, preferred integrations, infrastructure choices, and marketplace relationships all start reinforcing one another.

You are not only buying capability. You are stepping into a commercial lane that has already decided which kinds of tools, models, and partners are easiest to adopt.

This is where an early strategic question starts showing up for anyone outside the centre of that market. A lot of the modern AI stack still reflects American platform gravity:

  • American infrastructure.
  • American billing models.
  • American ecosystem partners.

Sometimes that is just because the market formed there first. Sometimes it looks more deliberate. Sometimes it feels like modern gatekeeping dressed up as product maturity. Time will tell how much is design and how much is momentum, but the pattern is worth noticing early.

The moment you step outside that loop into Open Source and Locally Run spectrum – (Ollama & Qwen), you feel the gap immediately.

  • Not because open models are useless.
  • Because the product system around them is thinner.
  • Less is pre-decided.
  • Less is pre-integrated.
  • Less is commercially smoothed over for you.

That freedom is real, but so is the extra engineering load, and so is the operational tax that comes with stitching your own environment together.

So the difference is bigger than Xcode versus another editor. It is the difference between entering a polished commercial system and stepping into a flexible technical landscape where you have to earn coherence yourself.

Switching to Flexibility: VS Code Trade-Off

Visual Studio Code opens everything up. That is its strength and its trap.

In Xcode:

  • The experience is cohesive because so many decisions are made on your behalf.

In VS Code:

  • The power comes from the opposite design philosophy.
  • You get modularity.
  • You get extensions.
  • You get cross-platform flexibility.
  • You get a far larger surface area for experimentation.

But the moment you have that freedom, you also inherit the burden of choice.

Closed product systems tend to win on day-one usability because they abstract the right engineering decisions at the right level. They do not remove architecture. They narrow the path. That is why they often feel closer to low-code than many engineers want to admit. You still need technical judgment, but you are not rebuilding the runway before the plane can move. For teams that care about speed, onboarding, and predictable execution, that matters.

Open systems are different. They play better with everything. They let you mix vendors, models, workflows, runtimes, and operating systems. They are more adaptable to edge cases and usually more resilient to vendor lock-in.

But they demand discipline much earlier. Someone has to:

  • Decide the stack.
  • Document it.
  • Train the next person.
  • Own the operational nuance once the first clever setup meets a second user, a production deadline, or a real support issue.

That is why the locking changes shape.

  • In a closed system, you are often locked into the vendor.
  • In an open one, you can very quickly become locked into your own key people, your own scripts, your own half-documented setup, and your own process debt.

Different lock-in. Same risk if you are not honest about it.

So when people say Xcode is polished and VS Code is powerful, that is true but incomplete. The real difference is that one packages engineering viability into the product experience, while the other gives you the pieces and assumes your team has the maturity to turn them into a system.

The Open Source Reality: You Have to Earn It

This is where the narrative around “just run it locally” starts to fall apart. Open source is not just a cheaper substitute for frontier systems. It is a different operating model.

The moment you leave the managed path, you are no longer simply consuming intelligence. You are assembling a stack.

That stack sounds simple in theory:

  • Install Ollama.
  • Pull a model.
  • Point your editor or workflow at it.

In practice, that simplicity disappears fast. You start making decisions about:

  • Which model family behaves best for coding.
  • Which quantisation gives you acceptable speed without collapsing quality.
  • How much context you can really afford.
  • How aggressively your machine throttles under load.
  • Whether the setup that worked yesterday still works after the next upgrade.

The technical decisions are only half the story. The other half is operational.

The more flexible the system becomes, the more maintenance migrates onto you. Updates, regressions, incompatible tooling, model swaps, broken integrations, and environment drift all become part of the cost base.

Nothing is charging you a monthly subscription for that directly, but you are still paying.

  • You are paying in time.
  • You are paying in attention.
  • You are paying in organisational complexity.

The Complexity Curve: Local Is Not Just “Free Compute”

Let’s be honest about the effort curve. Local is not free compute. It is deferred complexity.

Instead of paying for a polished service layer upfront, you absorb complexity across:

  • Model choice.
  • Runtime behaviour.
  • Hardware limits.
  • Heat.
  • Memory pressure.
  • Context management.
  • Version drift.
  • Workflow design.

What makes this harder is that the choices compound:

  • Pick a small model and you may get speed but lose quality.
  • Pick a larger one and your hardware becomes the bottleneck.
  • Push context and latency climbs.
  • Quantise aggressively and capability drops in ways that are not always obvious until you hit real work.

None of these are abstract benchmarks when you are trying to ship. They become practical trade-offs that shape how much trust you can place in the output.

This is why the public learning curve around local AI feels so much messier than the learning curve around a closed product. In a managed system, thousands of people are usually learning a similar path. In the local world, everyone is mixing slightly different hardware, slightly different model sizes, slightly different inference backends, and slightly different workflow assumptions.

The community is real and often generous, but the knowledge base is more fragmented because the permutations are far wider.

So the complexity is not just technical detail for its own sake. It is decision density. You are making more choices earlier, with less standardisation around you, while still trying to get useful work done. That is why local setups reward engineers who like system design and punish teams that expected a quick substitute for frontier convenience.

Local Reality

To get something meaningful running, you are usually doing more than installing software. You are:

  • Testing model sizes.
  • Evaluating quantisation trade-offs.
  • Tuning performance against memory headroom.
  • Watching thermal behaviour.
  • Deciding what quality loss you are willing to tolerate for speed.

That is not a side quest. That is the job.

By the time you have something stable, you have effectively started becoming your own infrastructure team. Not at data-centre scale, but at a very real workstation scale where bad choices still cost hours and good choices still need to be repeatable.

Contrast: Cloud Simplicity

With premium cloud models, the simplicity is not fake. It is engineered.

  • You sign in.
  • You choose the model.
  • You start working.

That sounds trivial until you compare how much hidden optimisation sits underneath that experience. The provider is handling scaling, infrastructure tuning, model refreshes, context management improvements, uptime, and a lot of the ugly work required to make the system feel fast and reliable.

You are also inheriting shared learning. Frontier platforms attract large user bases, which means patterns emerge faster:

  • Workflows get documented.
  • Problems get discussed publicly.
  • Tutorials accumulate.
  • Integrations mature.

If something breaks, odds are high that someone else has already hit it and found a workaround. That matters in the real world because community maturity lowers friction in ways benchmarks do not capture.

Cloud simplicity therefore is not just about convenience. It is about compressing operational uncertainty. You are paying for the model, yes, but you are also paying for a constantly maintained service wrapper around the model. For many teams, especially those trying to move quickly, that wrapper is a major part of the value.

Community vs Isolation

This is another under-discussed difference. Closed and frontier ecosystems benefit from concentration. The more people use the same tools in roughly similar ways, the faster the public body of knowledge improves.

  • Documentation gets better.
  • Integration examples become easier to find.
  • Best practice is easier to distinguish from random experimentation.

That does not eliminate bad advice, but it increases the chance that practical answers are already out there.

Local and open ecosystems are more fragmented by nature.

  • Different hardware.
  • Different editors.
  • Different runtimes.
  • Different model families.
  • Different quality expectations.

That creates freedom, but it also means public learning is spread across forums, GitHub issues, Discord threads, scattered blog posts, and highly context-specific fixes. You are often piecing together partial truths instead of following one mature operational path.

For experienced engineers, that fragmentation can be energising. For teams trying to scale capability across multiple people, it can become a process problem very quickly. Enablement, onboarding, documentation, and support all become part of the real-world cost equation. This is where human systems start mattering as much as model systems.

The Hardware Illusion: Bigger Machine ≠ Linear Gains

There’s a natural instinct:

“If I just get a bigger machine, this solves itself.”

The Reality: Even with a high-end dev kit (e.g. 128GB RAM and a top shelf GPU):

  • Performance improves… but not linearly.
  • You are still constrained by model optimisation.
  • You are still limited by inference speed.
  • You are still not matching hyperscaler efficiency.

And this becomes clear quickly: Hardware upgrades improve experience… but don’t eliminate architectural constraints.

The Supply Chain Signal: Where Optimisation Actually Happens

Then you zoom out and something more strategic appears. The global chip supply chain is not really optimising for individual developers trying to run frontier-adjacent workloads from a desk. It is optimising for the buyers that can absorb huge volumes of high-margin compute: hyperscalers, data-centre operators, and enterprise demand at scale.

That has consequences downstream:

  • Consumer hardware gets better every year, especially with integrated memory architectures and more capable SoCs.
  • But it still lives in the shadow of where the most aggressive optimisation effort is being spent.
  • That is why the local story improves, yet rarely catches the frontier story on equal terms.

So:

  • NVIDIA GPUs in terms of manufacturing quality and client are prioritised for data centres, not individual developers.
  • High VRAM configurations stay scarce and expensive because enterprise demand soaks up the best configurations.
  • Consumer hardware continues to trail enterprise capability in both memory and throughput.
  • And chip design is increasingly shifting toward tightly integrated SoC architectures that optimise whole-system efficiency rather than empowering endlessly modular DIY compute rigs.

Windows + NVIDIA Contrast

On paper, a Windows + NVIDIA setup should dominate. In practice:

  • GPU VRAM becomes the bottleneck.
  • High VRAM cards are expensive and scarce.
  • Consumer setups hit limits quickly.

So while:

  • CUDA acceleration is powerful.
  • Raw GPU compute is strong.

You still face Memory constraints and cost barriers at scale.

The Economic Layer: Where the Real Game Is Played

This is where everything becomes real. Not feature lists. Not vibe. Not benchmark screenshots from ideal conditions. Total cost over time, under actual usage, with real people, real deadlines, and real workflow friction.

The first misleading comparison is subscription versus hardware as if they are clean equivalents. They are not.

  • A subscription buys access to a moving target.
  • The hardware purchase buys a fixed ceiling that you then try to optimise around.
  • One improves underneath you.
  • The other depreciates underneath you.

Consumer subscription pricing looks cheap at first glance:

  • ChatGPT Plus is priced at twenty dollars a month, and higher-end plans move up from there.
  • Claude has similar consumer tiers.
  • API usage introduces token-based pricing that scales with workload rather than seat count.

On paper that can look expensive if you compare it to a once-off machine purchase. In practice the equation changes as soon as throughput, waiting time, and iteration speed become part of the analysis.

Now compare that to local ownership. A Mac Studio-class machine with memory high enough to make local model work genuinely interesting is not impulse-buy territory. Apple’s current Mac Studio line starts lower, but meaningful high-memory configurations push materially upward, and even Apple’s own product positioning makes clear that the bigger memory tiers are the ones aimed at heavyweight pro workflows rather than casual experimentation.

The Five-Year TCO Logic

This is where decision makers need to slow down and think properly. If you buy a high-memory machine outright, spread the capital cost over five years, and add electricity plus some maintenance friction, local can look compelling for predictable batch-heavy workloads.

But that only holds if:

  • The hardware is good enough for the work you need.
  • The team can support the stack.
  • The value of waiting is low.

The moment speed changes business value, the maths starts moving back toward cloud.

Frontier systems have another advantage that does not show up in simplistic TCO comparisons: annual platform improvement. The subscription is not only buying access for this month. It is buying access to:

  • The next model refresh.
  • The next optimiser pass.
  • The next infrastructure gain.
  • The next context improvement.
  • The next agent workflow upgrade.

Over five years, that compounds. Local hardware does not compound in the same way. It ages.

That is why the real threshold question is not ‘is local cheaper’ but ‘under what workload does local become cheaper enough to justify the slower operating model and heavier engineering burden’.

  • If the work is repetitive, asynchronous, cost-sensitive, and tolerant of lower frontier performance, local starts making sense.
  • If the work depends on rapid iteration, interactive debugging, collaborative experimentation, or high-stakes output quality, the cloud premium can easily pay for itself.

You can also see the split in token economics. API pricing can be brutally efficient for disciplined workflows and unexpectedly expensive for chaotic ones. Token-heavy exploratory work, large contexts, repeated rewrites, and long agent loops can push usage up fast. But the counterfactual matters. If a faster frontier loop saves engineer time, accelerates delivery, or reduces expensive mistakes, the token bill may still be cheaper than the hidden labour cost of making a local stack behave.

So the trade is not frontier versus local in the abstract. It is frontier subscription velocity versus local ownership control.

  • One optimises time.
  • The other can optimise unit cost once the workflow is stable enough.

Different workloads cross that threshold at different points, which is why dogmatic answers here are usually a red flag.

The First Misleading Comparison

Cloud looks cheap:

  • $20 – $100/month.
  • But that’s not real usage.

Because AI scales with curiosity, iteration, exploration, and team usage, and suddenly:

  • More prompts.
  • More tokens.
  • More cost.

The Hardware Counterpoint

  • Local: High upfront cost.
  • Lower marginal cost.
  • But: Time cost increases.
  • Energy cost appears.
  • Maintenance cost grows.

Real Example: Time vs Cost vs Energy

FeatureCloud (Codex)Local (Qwen via Ollama)
TaskCode review / Same analysisSame analysis
Time~5 minutes3 hours (and still going)
OutputStructured, immediate feedbackMachine maxed out
HardwareApple M3 Max (36GB RAM)Apple M3 Max (36GB RAM)
System LoadHigh, but AI compute offloaded.GPU and RAM maxed out, fans reminding me they exist.

The Creative Loop Trade – Off

This is where cloud dominates. When you are:

  • In a flow state.
  • Iterating quickly.
  • Testing ideas.

You need:

  • Low latency.
  • Fast turnaround.

Local breaks that rhythm.

But Local Introduces a New Pattern

Async workflows. You can:

  • Queue tasks overnight.
  • Run heavy analysis in batches.
  • Process large contexts slowly.

The Hybrid Insight Real-time vs asynchronous intelligence.

ModeUse Case
Real – Time (Cloud)Ideation, Rapid iteration, Creative work
Async (Local)Deep analysis, Batch processing, Cost – sensitive workloads

The Threshold Question

When does local win?

On paper:

  • High usage.
  • Repetitive tasks.
  • Stable workflows.

But Reality Adds Complexity You must factor:

  • Time.
  • Energy.
  • Opportunity cost.

The Hidden Shift

As workflows mature, humans become the latency, not the AI. If a local run takes two hours but your review, thinking, code inspection, and decision-making would have taken longer anyway, the gap starts to shrink. At that point the benchmark number stops telling the full story. What matters is whether the system can overlap useful work instead of forcing dead time.

That is where local starts becoming more interesting than the headline speed comparisons suggest. A developer can:

  • Review another module.
  • Clean up architecture.
  • Write documentation.
  • Assess previous output while a heavier local job runs in the background.

The raw output latency still matters, but it matters less once the workflow is designed around overlap rather than instant response.

This is also where the economics mature. Early in the journey, frontier systems win hard because speed unlocks learning, momentum, and fast iteration. Later, when patterns stabilise and the work becomes more batchable, local can start clawing back economic ground because the human is no longer sitting idle waiting for every answer.

Different users. Different workloads. Different opportunities. Different realities.

So the hidden shift is simple: Once you stop treating AI like a chatbot and start treating it like part of an operating model, the question changes from ‘how fast is the model’ to ‘how well does this workflow use human time’. That is a much more executive question than most benchmark debates ever become.

Counting the Intangibles (With Full Honesty)

Cloud

  • You pay for: Speed, Flow, Simplicity, Shared knowledge.

Local

  • You invest in: Capability, Control, Independence, Understanding.

The Parallel Nobody Wants to Admit

AI didn’t remove process. It renamed it.

  • Guardrails = governance
  • Context = architecture and documentation

Human vs AI

  • AI: Fast, Scalable, Pattern – driven.
  • Human: Context – aware, Strategic, Interpretive.

The Real Advantage Human judgment + AI execution.

Two Weeks In: The Honest Answer

Yes, it works. Yes, it can be commercialised. But:

  • Only with discipline.
  • Only with structure.
  • Only with awareness.

Where This Lands

  1. AI Didn’t Replace Developers: It shifted them.
  2. Economics Will Decide Adoption: Not hype.
  3. Hybrid Is a viable End State: With scale and time.

Final Thought

AI didn’t remove complexity. It exposed it. Faster.

AI is not the advantage. Structure is.

And that’s where the real separation happens next.

Discover more from The KeegzVerse

Subscribe now to keep reading and get access to the full archive.

Continue reading