<?xml version="1.0" encoding="utf-8" standalone="yes"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom">
  <channel>
    <title>Running Local AI on Intel Lunar Lake on Rosen Papazov</title>
    <link>https://shadowcp.dev/series/running-local-ai-on-intel-lunar-lake/</link>
    <description>Recent content in Running Local AI on Intel Lunar Lake on Rosen Papazov</description>
    <generator>Hugo -- gohugo.io</generator>
    <language>en</language>
    <copyright>&lt;a href=&#34;https://creativecommons.org/licenses/by-nc/4.0/&#34; target=&#34;_blank&#34; rel=&#34;noopener&#34;&gt;CC BY-NC 4.0&lt;/a&gt;</copyright>
    <lastBuildDate>Mon, 06 Apr 2026 12:00:00 +0300</lastBuildDate><atom:link href="https://shadowcp.dev/series/running-local-ai-on-intel-lunar-lake/index.xml" rel="self" type="application/rss+xml" />
    <item>
      <title>Running Local AI on Intel Lunar Lake — Part 1: Hardware, Drivers, and the Intel Compute Stack</title>
      <link>https://shadowcp.dev/posts/2026/04/running-local-ai-on-intel-lunar-lake-part-1-hardware-drivers-and-the-intel-compute-stack/</link>
      <pubDate>Mon, 06 Apr 2026 12:00:00 +0300</pubDate>
      
      <guid>https://shadowcp.dev/posts/2026/04/running-local-ai-on-intel-lunar-lake-part-1-hardware-drivers-and-the-intel-compute-stack/</guid>
      <description>Part 1 of 4 in a series on building a fully local AI development environment on an Intel Core Ultra 7 268V laptop running Fedora 43.
Why local AI? Cloud-hosted AI is convenient until it isn&amp;rsquo;t. Latency spikes, privacy concerns with proprietary code, rate limits during crunch time, and the monthly bill that creeps up. Running models locally solves all of these — if your hardware can handle it.
Intel&amp;rsquo;s Lunar Lake processors are interesting for this.</description>
    </item>
    
  </channel>
</rss>
