<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Ollama on Ben's Info Tech Blog</title><link>https://infotechwithben.com/tags/ollama/</link><description>Recent content in Ollama on Ben's Info Tech Blog</description><generator>Hugo</generator><language>en</language><lastBuildDate>Sat, 28 Mar 2026 00:00:00 +0000</lastBuildDate><atom:link href="https://infotechwithben.com/tags/ollama/index.xml" rel="self" type="application/rss+xml"/><item><title>Shadow Stack — Running Local AI Models on Consumer Hardware</title><link>https://infotechwithben.com/posts/shadow-stack-complete-guide/</link><pubDate>Sat, 28 Mar 2026 00:00:00 +0000</pubDate><guid>https://infotechwithben.com/posts/shadow-stack-complete-guide/</guid><description>What Is the Shadow Stack? The &amp;ldquo;shadow stack&amp;rdquo; is a local inference layer that runs alongside your cloud API usage. Instead of every prompt hitting OpenAI or Anthropic, lightweight or private workloads run on GPUs you already own. You choose the right tier per task.
Three deployment tiers:
Cloud APIs — Claude, GPT-4o, Gemini. Highest quality, per-token cost, zero ops. Local inference — Llama 3, Mistral, Phi-3 on your hardware. Fixed cost after setup, full data sovereignty.</description></item></channel></rss>