<?xml version="1.0" encoding="utf-8" standalone="yes"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:content="http://purl.org/rss/1.0/modules/content/">
  <channel>
    <title>Llm on smcleod.net</title>
    <link>https://smcleod.net/tags/llm/</link>
    <description>Recent content in Llm on smcleod.net</description>
    <image>
      <title>smcleod.net</title>
      <url>https://smcleod.net/logo-painted-350.svg</url>
      <link>https://smcleod.net/logo-painted-350.svg</link>
    </image>
    <generator>Hugo -- 0.152.2</generator>
    <language>en-GB</language>
    <lastBuildDate>Thu, 02 Apr 2026 14:12:42 +1100</lastBuildDate>
    <atom:link href="https://smcleod.net/tags/llm/index.xml" rel="self" type="application/rss+xml" />
    <item>
      <title>Patching NVIDIA&#39;s driver and vLLM to enable P2P on consumer GPUs</title>
      <link>https://smcleod.net/2026/02/patching-nvidias-driver-and-vllm-to-enable-p2p-on-consumer-gpus/</link>
      <pubDate>Wed, 25 Feb 2026 23:59:00 +1000</pubDate>
      <guid>https://smcleod.net/2026/02/patching-nvidias-driver-and-vllm-to-enable-p2p-on-consumer-gpus/</guid>
      <description>NVIDIA artificially restricts peer-to-peer (P2P) GPU communication to their enterprise cards. Turns out this is a software limitation, not a hardware one. I patched my drivers to remove it, hacked vLLM to take advantage of it, and got a 15-50% throughput improvement running Qwen 3.5 35b on dual RTX 3090s.</description>
    </item>
    <item>
      <title>The Role Bridging Problem</title>
      <link>https://smcleod.net/2025/11/the-role-bridging-problem/</link>
      <pubDate>Thu, 20 Nov 2025 01:10:00 +1000</pubDate>
      <guid>https://smcleod.net/2025/11/the-role-bridging-problem/</guid>
      <description>An observation on functional correctness without domain quality.</description>
    </item>
    <item>
      <title>Stop Polluting Context - Let Users Disable Individual MCP Tools</title>
      <link>https://smcleod.net/2025/08/stop-polluting-context-let-users-disable-individual-mcp-tools/</link>
      <pubDate>Tue, 26 Aug 2025 01:10:00 +1000</pubDate>
      <guid>https://smcleod.net/2025/08/stop-polluting-context-let-users-disable-individual-mcp-tools/</guid>
      <description>If you&amp;rsquo;re building MCP servers, you should be adding the ability to disable individual tools.</description>
    </item>
    <item>
      <title>MCP DevTools</title>
      <link>https://smcleod.net/2025/08/mcp-devtools/</link>
      <pubDate>Mon, 18 Aug 2025 01:10:00 +1000</pubDate>
      <guid>https://smcleod.net/2025/08/mcp-devtools/</guid>
      <description>A single, modular MCP server for AI coding agents.</description>
    </item>
    <item>
      <title>Agentic Coding Adoption Cost Cycle</title>
      <link>https://smcleod.net/2025/07/agentic-coding-adoption-cost-cycle/</link>
      <pubDate>Wed, 16 Jul 2025 01:10:00 +1000</pubDate>
      <guid>https://smcleod.net/2025/07/agentic-coding-adoption-cost-cycle/</guid>
      <description>The two common themes I see with engineers adopting agentic coding tools</description>
    </item>
    <item>
      <title>Agentic Coding Workflow &amp; Cline Demo</title>
      <link>https://smcleod.net/2025/06/agentic-coding-workflow-cline-demo/</link>
      <pubDate>Fri, 20 Jun 2025 01:10:00 +1000</pubDate>
      <guid>https://smcleod.net/2025/06/agentic-coding-workflow-cline-demo/</guid>
      <description>Recording of a demo and Q&amp;amp;A session on my Agentic Coding workflow and Cline use.</description>
    </item>
    <item>
      <title>Vibe Coding vs Agentic Coding</title>
      <link>https://smcleod.net/2025/06/vibe-coding-vs-agentic-coding/</link>
      <pubDate>Fri, 06 Jun 2025 01:10:00 +1000</pubDate>
      <guid>https://smcleod.net/2025/06/vibe-coding-vs-agentic-coding/</guid>
      <description>From Creative Exploration to Production Quality</description>
    </item>
    <item>
      <title>My Plan, Document, Act, Review flow for Agentic Software Development</title>
      <link>https://smcleod.net/2025/04/my-plan-document-act-review-flow-for-agentic-software-development/</link>
      <pubDate>Mon, 28 Apr 2025 01:10:00 +1000</pubDate>
      <guid>https://smcleod.net/2025/04/my-plan-document-act-review-flow-for-agentic-software-development/</guid>
      <description>An overview of my agentic coding development flow.</description>
    </item>
    <item>
      <title>LLM Sampling Parameters Guide</title>
      <link>https://smcleod.net/2025/04/llm-sampling-parameters-guide/</link>
      <pubDate>Fri, 25 Apr 2025 01:10:00 +1000</pubDate>
      <guid>https://smcleod.net/2025/04/llm-sampling-parameters-guide/</guid>
      <description>A practical guide to LLM sampling parameters for Ollama, llama.cpp, and MLX</description>
    </item>
    <item>
      <title>Getting Started with Agentic Systems - Developer Learning Paths</title>
      <link>https://smcleod.net/2025/04/getting-started-with-agentic-systems-developer-learning-paths/</link>
      <pubDate>Wed, 16 Apr 2025 01:10:00 +1000</pubDate>
      <guid>https://smcleod.net/2025/04/getting-started-with-agentic-systems-developer-learning-paths/</guid>
      <description>A curated learning path for engineers looking to gain practical experience with AI and agentic systems.</description>
    </item>
    <item>
      <title>The Cost of Agentic Coding</title>
      <link>https://smcleod.net/2025/04/the-cost-of-agentic-coding/</link>
      <pubDate>Thu, 03 Apr 2025 01:00:01 +1000</pubDate>
      <guid>https://smcleod.net/2025/04/the-cost-of-agentic-coding/</guid>
      <description>The cost of not leveraging agentic coding is likely far greater than you think.</description>
    </item>
    <item>
      <title>The Democratisation Paradox: What History Teaches Us About AI</title>
      <link>https://smcleod.net/2025/03/the-democratisation-paradox-what-history-teaches-us-about-ai/</link>
      <pubDate>Sun, 23 Mar 2025 01:00:01 +1000</pubDate>
      <guid>https://smcleod.net/2025/03/the-democratisation-paradox-what-history-teaches-us-about-ai/</guid>
      <description>The Democratisation Paradox: What History Teaches Us About AI.</description>
    </item>
    <item>
      <title>The effects of prompt caching on Agentic coding</title>
      <link>https://smcleod.net/2025/03/the-effects-of-prompt-caching-on-agentic-coding/</link>
      <pubDate>Thu, 20 Mar 2025 21:00:01 +1000</pubDate>
      <guid>https://smcleod.net/2025/03/the-effects-of-prompt-caching-on-agentic-coding/</guid>
      <description>What is prompt caching and why is it so important for Agentic coding?</description>
    </item>
    <item>
      <title>Agentic Coding - Live Demo / Brownbag</title>
      <link>https://smcleod.net/2025/02/agentic-coding-live-demo-/-brownbag/</link>
      <pubDate>Fri, 07 Feb 2025 01:00:01 +1000</pubDate>
      <guid>https://smcleod.net/2025/02/agentic-coding-live-demo-/-brownbag/</guid>
      <description>Video recording of a brownbag presentation / live demo I ran on Agentic Coding using Cline/Roo Code</description>
    </item>
    <item>
      <title>Bringing K/V Context Quantisation to Ollama</title>
      <link>https://smcleod.net/2024/12/bringing-k/v-context-quantisation-to-ollama/</link>
      <pubDate>Wed, 04 Dec 2024 20:00:02 +0000</pubDate>
      <guid>https://smcleod.net/2024/12/bringing-k/v-context-quantisation-to-ollama/</guid>
      <description>K/V context cache quantisation has been added to Ollama. This enables significant reductions in VRAM usage, allowing users to realise the potential of expanded context sizes and run larger models at their existing context sizes.</description>
    </item>
    <item>
      <title>Will AI Take My Job?</title>
      <link>https://smcleod.net/2024/11/will-ai-take-my-job/</link>
      <pubDate>Fri, 29 Nov 2024 22:00:02 +0000</pubDate>
      <guid>https://smcleod.net/2024/11/will-ai-take-my-job/</guid>
      <description>It&amp;#39;s probably not so much AI itself you have to fear.</description>
    </item>
    <item>
      <title>Generating Diagrams with with AI / LLMs</title>
      <link>https://smcleod.net/2024/10/generating-diagrams-with-with-ai-/-llms/</link>
      <pubDate>Tue, 08 Oct 2024 01:00:10 +0000</pubDate>
      <guid>https://smcleod.net/2024/10/generating-diagrams-with-with-ai-/-llms/</guid>
      <description>Generating diagrams with AI / LLMs</description>
    </item>
    <item>
      <title>Ingest: Streamlining Content Preparation for LLMs</title>
      <link>https://smcleod.net/2024/07/ingest-streamlining-content-preparation-for-llms/</link>
      <pubDate>Mon, 29 Jul 2024 01:00:10 +0000</pubDate>
      <guid>https://smcleod.net/2024/07/ingest-streamlining-content-preparation-for-llms/</guid>
      <description>A CLI tool for parsing directories into LLM-friendly markdown</description>
    </item>
    <item>
      <title>LLM Parameter Playground</title>
      <link>https://smcleod.net/2024/07/llm-parameter-playground/</link>
      <pubDate>Sat, 20 Jul 2024 01:00:10 +0000</pubDate>
      <guid>https://smcleod.net/2024/07/llm-parameter-playground/</guid>
      <description>A web app to explore the effects of different inference parameters on LLMs.</description>
    </item>
    <item>
      <title>Code, Chaos, and Copilots (AI/LLM Talk July 2024)</title>
      <link>https://smcleod.net/2024/07/code-chaos-and-copilots-ai/llm-talk-july-2024/</link>
      <pubDate>Thu, 18 Jul 2024 01:00:10 +0000</pubDate>
      <guid>https://smcleod.net/2024/07/code-chaos-and-copilots-ai/llm-talk-july-2024/</guid>
      <description>Integrating AI into Daily Development</description>
    </item>
    <item>
      <title>Understanding AI/LLM Quantisation Through Interactive Visualisations</title>
      <link>https://smcleod.net/2024/07/understanding-ai/llm-quantisation-through-interactive-visualisations/</link>
      <pubDate>Wed, 17 Jul 2024 01:00:10 +0000</pubDate>
      <guid>https://smcleod.net/2024/07/understanding-ai/llm-quantisation-through-interactive-visualisations/</guid>
      <description>AI/LLM Quantisation Visualised</description>
    </item>
    <item>
      <title>Gollama: Ollama Model Manager</title>
      <link>https://smcleod.net/2024/06/gollama-ollama-model-manager/</link>
      <pubDate>Sat, 01 Jun 2024 01:00:10 +0000</pubDate>
      <guid>https://smcleod.net/2024/06/gollama-ollama-model-manager/</guid>
      <description>Gollama is a Go-based client for Ollama for managing models.</description>
    </item>
    <item>
      <title>Confuddlement: Download Confluence Spaces as Markdown, Summarise with Ollama</title>
      <link>https://smcleod.net/2024/05/confuddlement-download-confluence-spaces-as-markdown-summarise-with-ollama/</link>
      <pubDate>Thu, 23 May 2024 05:45:10 +0000</pubDate>
      <guid>https://smcleod.net/2024/05/confuddlement-download-confluence-spaces-as-markdown-summarise-with-ollama/</guid>
      <description>Download Confluence Spaces as Markdown, Summarise with Ollama</description>
    </item>
    <item>
      <title>SuperPrompter - Supercharge your text prompts for AI/LLM image generation</title>
      <link>https://smcleod.net/2024/03/superprompter-supercharge-your-text-prompts-for-ai/llm-image-generation/</link>
      <pubDate>Fri, 22 Mar 2024 00:00:00 +0000</pubDate>
      <guid>https://smcleod.net/2024/03/superprompter-supercharge-your-text-prompts-for-ai/llm-image-generation/</guid>
      <description>&lt;p&gt;&lt;a href=&#34;https://github.com/sammcj/superprompter&#34;&gt;SuperPrompter&lt;/a&gt; is a Python-based application that utilises the &lt;a href=&#34;https://huggingface.co/roborovski/superprompt-v1&#34;&gt;SuperPrompt-v1&lt;/a&gt; model to generate optimised text prompts for AI/LLM image generation (for use with Stable Diffusion etc&amp;hellip;) from user prompts.&lt;/p&gt;
&lt;p&gt;&lt;img alt=&#34;screenshot&#34; loading=&#34;lazy&#34; src=&#34;https://smcleod.net/2024/03/superprompter-supercharge-your-text-prompts-for-ai/llm-image-generation/superprompter-screenshot.jpg&#34;&gt;&lt;/p&gt;
&lt;p&gt;See &lt;a href=&#34;https://brianfitzgerald.xyz/prompt-augmentation/&#34;&gt;Brian Fitzgerald&amp;rsquo;s Blog&lt;/a&gt; for a detailed explanation of the SuperPrompt-v1 model and its capabilities / limitations.&lt;/p&gt;
&lt;!-- more --&gt;
&lt;h2 id=&#34;features&#34;&gt;Features&lt;/h2&gt;
&lt;ul&gt;
&lt;li&gt;Utilises the &lt;a href=&#34;https://huggingface.co/roborovski/superprompt-v1&#34;&gt;SuperPrompt-v1&lt;/a&gt; model for text generation.&lt;/li&gt;
&lt;li&gt;A basic (aka ugly) graphical user interface built with tkinter.&lt;/li&gt;
&lt;li&gt;Customisable generation parameters (max new tokens, repetition penalty, temperature, top p, top k, seed).&lt;/li&gt;
&lt;li&gt;Optional logging of input parameters and generated outputs.&lt;/li&gt;
&lt;li&gt;Bundling options to include or exclude pre-downloaded model files.&lt;/li&gt;
&lt;li&gt;Unloads the models when the application is idle to free up memory.&lt;/li&gt;
&lt;/ul&gt;
&lt;h2 id=&#34;prebuilt-binaries&#34;&gt;Prebuilt Binaries&lt;/h2&gt;
&lt;p&gt;Check &lt;a href=&#34;https://github.com/sammcj/superprompter/releases&#34;&gt;releases&lt;/a&gt; page to see if there are any prebuilt binaries available for your platform.&lt;/p&gt;</description>
    </item>
    <item>
      <title>Llamalink - Ollama to LM Studio LLM Model Linker</title>
      <link>https://smcleod.net/2024/03/llamalink-ollama-to-lm-studio-llm-model-linker/</link>
      <pubDate>Thu, 21 Mar 2024 00:00:00 +0000</pubDate>
      <guid>https://smcleod.net/2024/03/llamalink-ollama-to-lm-studio-llm-model-linker/</guid>
      <description>This is a simple command-line tool that allows you to easily link Ollama models to LM Studio&amp;#39;s directory structure. It simplifies the process of symlinking Ollama models to LM Studio, making it convenient to use the models in both applications.</description>
    </item>
    <item>
      <title>Open source, locally hosted AI powered Siri replacement</title>
      <link>https://smcleod.net/2023/11/open-source-locally-hosted-ai-powered-siri-replacement/</link>
      <pubDate>Mon, 20 Nov 2023 01:00:00 +0000</pubDate>
      <guid>https://smcleod.net/2023/11/open-source-locally-hosted-ai-powered-siri-replacement/</guid>
      <description>Using an ESP32 S3 Box 3 with Willow, Home Assistant and Large Language Models (LLMs) create a locally hosted, offline, AI powered Siri / Alexa replacement.</description>
    </item>
    <item>
      <title>Introduction to AI and Large Language Models (LLMs)</title>
      <link>https://smcleod.net/2023/01/introduction-to-ai-and-large-language-models-llms/</link>
      <pubDate>Thu, 26 Jan 2023 01:00:00 +0000</pubDate>
      <guid>https://smcleod.net/2023/01/introduction-to-ai-and-large-language-models-llms/</guid>
      <description>A high level intro to LLMs that I&amp;#39;m writing for a few friends that are new to the concept. It is far from complete, definitely contains some errors and is a work in progress.</description>
    </item>
    <item>
      <title>LLM FAQ</title>
      <link>https://smcleod.net/llm-faq/</link>
      <pubDate>Mon, 01 Jan 0001 00:00:00 +0000</pubDate>
      <guid>https://smcleod.net/llm-faq/</guid>
      <description>Frequently Asked Questions about LLMs and AI</description>
    </item>
    <item>
      <title>LLM vRAM Estimator</title>
      <link>https://smcleod.net/vram-estimator/</link>
      <pubDate>Mon, 01 Jan 0001 00:00:00 +0000</pubDate>
      <guid>https://smcleod.net/vram-estimator/</guid>
      <description>LLM vRAM Estimator</description>
    </item>
  </channel>
</rss>
