<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0"
     xmlns:atom="http://www.w3.org/2005/Atom"
     xmlns:content="http://purl.org/rss/1.0/modules/content/">
  <channel>
    <title>Will Angel's Blog</title>
    <link>https://www.williamangel.net/blog/</link>
    <description>Posts from williamangel.net</description>
    <language>en-us</language>
    <lastBuildDate>Thu, 07 May 2026 00:00:00 +0000</lastBuildDate>
    <atom:link href="https://www.williamangel.net/rss.xml" rel="self" type="application/rss+xml"/>
    <item>
      <title>Offline Agentic Coding: OpenCode</title>
      <link>https://www.williamangel.net/blog/2026/05/07/offline-agentic-coding-2.html</link>
      <guid isPermaLink="true">https://www.williamangel.net/blog/2026/05/07/offline-agentic-coding-2.html</guid>
      <pubDate>Thu, 07 May 2026 00:00:00 +0000</pubDate>
      <description>Limitations of offline coding</description>
      <content:encoded><![CDATA[<h1>Offline Agentic Coding part 2: OpenCode &amp; Kilocode.</h1>
<h3>Published 2026-05-07</h3>
<p><img src="2026-05-07-offline_agentic_coding_2.png" alt="offline agentic coding: a handdrawn clock"></p>
<h2>OpenCode:</h2>
<p>Claude code with non-anthropic models feels limited. Conveniently, we can also use OpenCode!</p>
<pre><code>ollama launch opencode</code></pre>
<p>OpenCode is like Claude Code but bring your own model.  Overall it&#x27;s very comparable, but is less polished in some ways while feeling more solid in others.</p>
<h2>Kilocode:</h2>
<p>Kilocode (kilo.ai) is an agentic coding platform. It&#x27;s now primarily powered by a fork of OpenCode. Their VScode/codium extension is very nice, and offers one of the my favorite views of the context window:</p>
<p><img src="2026-05-07-kilocode.png" alt="kilocode screenshot"></p>
<p>Hooking up a local model to the Kilocode vscode extension is straightforwad. Just give it your local server port and you&#x27;re good to go.</p>
<p>Kilocode is also nice because they have support for a wide range of model providers as well as their own model provider platform, so it&#x27;s seemless to switch between local, open, and proprietary models.</p>
<h2>Overall</h2>
<p>Overall local models for <strong>coding</strong> are still too slow to be practical on regular hardware, but it&#x27;s nice to have as a capability if the internet goes down, and it&#x27;s still magical to be able to tell me computer to program itself.</p>]]></content:encoded>
    </item>
    <item>
      <title>Offline Agentic Coding</title>
      <link>https://www.williamangel.net/blog/2026/04/27/offline-agentic-coding.html</link>
      <guid isPermaLink="true">https://www.williamangel.net/blog/2026/04/27/offline-agentic-coding.html</guid>
      <pubDate>Mon, 27 Apr 2026 00:00:00 +0000</pubDate>
      <description>Offline Agentic Coding: Ollama and Claude code</description>
      <content:encoded><![CDATA[<h1>Offline Agentic Coding</h1>
<h3>Published 2026-04-27</h3>
<p><img src="2026-04-27-offline_agentic_coding.png" alt="offline agentic coding: a handdrawn aeroplane"></p>
<p>You can use ollama as the backend for claude code!</p>
<pre><code>ollama launch claude --model</code></pre>
<p>This allows you to use claude code with local models. I&#x27;m writing this from an airplane with no internet connection.</p>
<h2>Overall model comparisons</h2>
<p>Gemma4:e2b did not finish any tasks despite being blazing fast at over 100 tokens per second.</p>
<p>qwen3-coder-next:q4_K_M actually did reasonably well.  Felt a bit worse than haiku quality but notably slower. Took around half an hour to fill up 75k of context, which is about 40 tokens per second while taking 50-60gb of memory.</p>
<p>qwen3.6:35b was also fairly reasonable. Did an adaquate job writing a small local data processing job, but was also fairly slow.</p>
<p>Gemma4:31b felt the most &#x27;claude-like&#x27; in claude code, but was also fairly slow and occasionally required some jostling and interruption.</p>
<h2>Overall</h2>
<p>I don&#x27;t seriously recommend local agentic coding with LLMs. You need some serious hardware to run decent models and it&#x27;s still slow. It&#x27;s a nice capability to have locally, but it probably isn&#x27;t better than coding by hand. Still very cool to have a computer that can program itself though, and amazing that a consumer device can locally run models and software that matches the original gpt-3 era ChatGPT style experience.</p>]]></content:encoded>
    </item>
    <item>
      <title>Washington DC on track for most volatile temperature year since 1959</title>
      <link>https://www.williamangel.net/blog/2026/04/19/Washington_DC_On_Track_For_Stormy_2026.html</link>
      <guid isPermaLink="true">https://www.williamangel.net/blog/2026/04/19/Washington_DC_On_Track_For_Stormy_2026.html</guid>
      <pubDate>Sun, 19 Apr 2026 00:00:00 +0000</pubDate>
      <description>An analysis of 85 years of daily weather data from Reagan National Airport.</description>
      <content:encoded><![CDATA[<h1>Washington DC on track for most volatile temperature year since 1959</h1>
<p>An analysis of 85 years of daily weather data from Reagan National Airport.</p>
<p><img src="Washington_DC_On_Track_For_Stormy_2026_figure2.png" alt="Temperature volatility chart"></p>]]></content:encoded>
    </item>
  </channel>
</rss>
