<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Local-Ai on Ben's Info Tech Blog</title><link>https://infotechwithben.com/tags/local-ai/</link><description>Recent content in Local-Ai on Ben's Info Tech Blog</description><generator>Hugo</generator><language>en</language><lastBuildDate>Sat, 28 Mar 2026 00:00:00 +0000</lastBuildDate><atom:link href="https://infotechwithben.com/tags/local-ai/index.xml" rel="self" type="application/rss+xml"/><item><title>Shadow Stack — Local vs API Cost Analysis</title><link>https://infotechwithben.com/posts/shadow-stack-local-vs-api-cost-analysis/</link><pubDate>Sat, 28 Mar 2026 00:00:00 +0000</pubDate><guid>https://infotechwithben.com/posts/shadow-stack-local-vs-api-cost-analysis/</guid><description>Hardware TCO Sample Build: RTX 3090 24GB (used market, ~$700–800) Component Cost RTX 3090 (used) $750 Host server (Proxmox, used workstation) $400 32 GB RAM $80 1 TB NVMe $80 Power (600W system × 8h/day × $0.12/kWh) ~$21/month One-time hardware ~$1,310 Monthly power ~$21 Break-even calculation — you need to compare against API costs you&amp;rsquo;d otherwise pay.
RTX 4060 Ti 16GB Build (~Budget) Component Cost RTX 4060 Ti 16GB $450 Mini PC / used workstation $200 16 GB RAM $40 512 GB NVMe $50 Power (250W × 8h × $0.</description></item><item><title>Shadow Stack — Running Local AI Models on Consumer Hardware</title><link>https://infotechwithben.com/posts/shadow-stack-complete-guide/</link><pubDate>Sat, 28 Mar 2026 00:00:00 +0000</pubDate><guid>https://infotechwithben.com/posts/shadow-stack-complete-guide/</guid><description>What Is the Shadow Stack? The &amp;ldquo;shadow stack&amp;rdquo; is a local inference layer that runs alongside your cloud API usage. Instead of every prompt hitting OpenAI or Anthropic, lightweight or private workloads run on GPUs you already own. You choose the right tier per task.
Three deployment tiers:
Cloud APIs — Claude, GPT-4o, Gemini. Highest quality, per-token cost, zero ops. Local inference — Llama 3, Mistral, Phi-3 on your hardware. Fixed cost after setup, full data sovereignty.</description></item></channel></rss>