<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:media="http://search.yahoo.com/mrss/"><channel><title><![CDATA[Tomcw.xyz]]></title><description><![CDATA[Writing about open infrastructure, data, technology, learning and a better world. From Tom C W.]]></description><link>https://tomcw.xyz/</link><image><url>https://tomcw.xyz/favicon.png</url><title>Tomcw.xyz</title><link>https://tomcw.xyz/</link></image><generator>Ghost 5.88</generator><lastBuildDate>Thu, 16 Apr 2026 20:39:20 GMT</lastBuildDate><atom:link href="https://tomcw.xyz/rss/" rel="self" type="application/rss+xml"/><ttl>60</ttl><item><title><![CDATA[Is This Just a Skill?]]></title><description><![CDATA[At one point everything was an app, or at least someone said it should be an app. Now I wonder, is everything a SKILL? And what's the difference between a skill and an actual product. I consider this with something I built and then decided...I think this is a skill. ]]></description><link>https://tomcw.xyz/is-this-just-a-skill/</link><guid isPermaLink="false">69c64d564e4f0e383144ca51</guid><category><![CDATA[ai]]></category><category><![CDATA[Tech]]></category><category><![CDATA[Systems]]></category><dc:creator><![CDATA[Tom Watson]]></dc:creator><pubDate>Fri, 27 Mar 2026 10:20:06 GMT</pubDate><media:content url="https://images.unsplash.com/photo-1679108230151-77401eb692b6?crop=entropy&amp;cs=tinysrgb&amp;fit=max&amp;fm=jpg&amp;ixid=M3wxMTc3M3wwfDF8c2VhcmNofDE4fHxza2lsbHxlbnwwfHx8fDE3NzQ2MDY0NjB8MA&amp;ixlib=rb-4.1.0&amp;q=80&amp;w=2000" medium="image"/><content:encoded><![CDATA[<img src="https://images.unsplash.com/photo-1679108230151-77401eb692b6?crop=entropy&amp;cs=tinysrgb&amp;fit=max&amp;fm=jpg&amp;ixid=M3wxMTc3M3wwfDF8c2VhcmNofDE4fHxza2lsbHxlbnwwfHx8fDE3NzQ2MDY0NjB8MA&amp;ixlib=rb-4.1.0&amp;q=80&amp;w=2000" alt="Is This Just a Skill?"><p>I built an app a few months ago. It&apos;s called <a href="https://impact-path.vercel.app/?ref=tomcw.xyz">ImpactPath</a>. You can sign up, log in, have a conversation with an AI that guides you through creating a theory of change for your organisation, and at the end you get professional outputs: a logic model, executive summary, visual diagram. It works. It&apos;s a genuine product if a little unfinished.</p><p>I stopped building it, partly because I need to make choices about where I spend my time and I think there are more interesting and valuable ways of helping people thinking about impact and learning. But also... I started wondering: is this just a skill?</p><h2 id="are-you-just-a-md-file">Are you just a .md file?</h2><p>There&apos;s a tongue-in-cheek website doing the rounds called <a href="https://deathbyclawd.com/?ref=tomcw.xyz">Death by Clawd</a> &#x2014; the &quot;SaaSpocalypse Survival Scanner.&quot; You enter a SaaS company and it rates the probability of being replaced by a Claude Skill. It&apos;s a joke, maybe, but within every joke, isn&apos;t there a semblance of truth?</p><p>A lot of software is essentially packaged expertise with a billing system attached. And if the expertise can be written down, if it&apos;s procedural knowledge, decision trees, best practices, structured processes, then maybe it can be a skill instead of a subscription.</p><p>Now that AI can produce things at speed, we can make so many more things.  The cost of building has collapsed, an idea in the morning can be a deployed app by evening.</p><p>But for many of those things, maybe most of them, should they be apps at all? Or are they just a SKILL.md file waiting to happen?</p><p>There&apos;s now an open standard for packaging expertise in a format that AI agents can use. It&apos;s called Agent Skills, and the specification lives at <a href="https://agentskills.io/?ref=tomcw.xyz">agentskills.io</a>. Originally developed by Anthropic, it&apos;s been adopted across Claude, OpenAI Codex, GitHub Copilot, VS Code, Cursor, and over twenty other platforms.</p><p>A skill is a folder containing instructions, scripts, and resources. The core is a SKILL.md file: metadata describing what it does and when to trigger, then markdown with the actual guidance. You can include reference files, templates, executable scripts. The agent loads what it needs, when it needs it.</p><p>So now, every time I start building something, I find myself asking: is this just a skill?</p><h2 id="what-i-actually-built-for-impactpath">What I actually built for ImpactPath</h2><p>Before I could build the app, I had to build the expertise layer:</p><ul><li>A data schema for capturing organisational context: mission, beneficiaries, activities, resources, geographic scope</li><li>A structured conversation flow that moves through stages: context gathering, activity mapping, outcome definition, assumption surfacing, measurement planning</li><li>Methodology for moving from vague impact language (&quot;we help people&quot;) to specific outcome statements (&quot;participants report increased confidence in managing their finances within 3 months&quot;)</li><li>Guidance on surfacing assumptions: what has to be true for this theory to hold?</li><li>Indicator selection logic: qualitative vs quantitative, primary vs secondary sources, realistic collection frequency</li><li>Output templates for different audiences: funder-facing, board-facing, community-facing</li></ul><p>The app wraps all this in authentication, saved progress, a nice interface, export buttons. But the real core of it is the methodology, the schema, the process, the examples. </p><p>And that could all live in a SKILL.md file.... in fact I DID extract all this into a skill <a href="https://github.com/dataforaction-tom/claude-code-template/blob/master/.claude/skills/theory-of-change-builder/SKILL.md?ref=tomcw.xyz" rel="noreferrer">which you can see here!</a></p><figure class="kg-card kg-bookmark-card"><a class="kg-bookmark-container" href="https://github.com/dataforaction-tom/claude-code-template/blob/master/.claude/skills/theory-of-change-builder/SKILL.md?ref=tomcw.xyz"><div class="kg-bookmark-content"><div class="kg-bookmark-title">claude-code-template/.claude/skills/theory-of-change-builder/SKILL.md at master &#xB7; dataforaction-tom/claude-code-template</div><div class="kg-bookmark-description">Ready-to-use project template for Claude Code with state tracking, slash commands, subagents, and continuous improvement - dataforaction-tom/claude-code-template</div><div class="kg-bookmark-metadata"><img class="kg-bookmark-icon" src="https://github.githubassets.com/assets/pinned-octocat-093da3e6fa40.svg" alt="Is This Just a Skill?"><span class="kg-bookmark-author">GitHub</span><span class="kg-bookmark-publisher">dataforaction-tom</span></div></div><div class="kg-bookmark-thumbnail"><img src="https://opengraph.githubassets.com/c864abe8744c19c7be29a770a1bdd0a343e6cad79e5453f246aa30d00ee8c0ff/dataforaction-tom/claude-code-template" alt="Is This Just a Skill?"></div></a></figure><h2 id="so-when-is-something-just-a-skill">So when is something just a skill?</h2><p>A skill is enough when the value is in the knowledge, not the infrastructure.</p><p>If what you&apos;re building is essentially &quot;here&apos;s how to do this thing well&quot;,  a methodology, a process, a set of decision rules, a framework for thinking through a problem that&apos;s probably a skill. The AI agent provides the interface. The skill provides the expertise.</p><p>Where does a creating a Theory of change sit on this? It&apos;s the expertise, the questions to ask, the order to ask them, the way to structure the outputs. A skill can encode all of that.</p><p>Brand guidelines? I have skills for The Good Ship brand.  Colours, fonts, layout patterns, tone of voice. From here I can create slides or documents or use it in HTML and CSS for websites</p><p>I think a skill is enough when:</p><ul><li>The value is procedural knowledge that can be written down</li><li>The user already has access to an AI agent</li><li>The outputs are documents, decisions, or structured data rather than ongoing services</li><li>There&apos;s no need to persist <strong>state</strong> across sessions (or the user&apos;s own systems handle that)</li><li>Security and access control aren&apos;t critical differentiators</li></ul><h2 id="when-something-goes-beyond-a-skill">When something goes beyond a skill</h2><p>But sometimes a skill isn&apos;t enough. Here&apos;s when you probably need more:</p><p><strong>Scale and concurrency.</strong> A skill runs in one person&apos;s conversation, but if you need to handle thousands of simultaneous users, coordinate between them, or maintain shared state, you need infrastructure. </p><p><strong>Observability and analytics.</strong> Skills are somewhat opaque, meaning you don&apos;t easily get aggregate data on how they&apos;re being used, where people get stuck, what outputs they&apos;re generating. If understanding usage patterns is core to the value proposition, for improvement, for reporting, for demonstrating impact, you probably need an app layer that captures this.</p><p><strong>Repeatability and consistency.</strong> Skills depend on the underlying model, which updates, changes, has different behaviours across platforms. If you need guaranteed consistent outputs, the same input always* produces the same output, for compliance or contractual reasons then you need to control more of the stack.</p><p><strong>Services, not just outputs.</strong> A skill produces outputs: documents, recommendations, structured data. But what if the value is ongoing service? Monitoring, alerting, scheduled tasks, integrations that run without human prompting. That&apos;s not a skill.</p><p><strong>Security and access control.</strong> Skills inherit the security model of wherever they run. If you need fine-grained access control, audit trails, data residency guarantees, or compliance certifications, you need infrastructure that provides these. A skill for handling sensitive health data is risky if you can&apos;t control where it runs.</p><p><strong>Identity and payment.</strong> If you need to know who&apos;s using it, limit access, or charge money, you need infrastructure around authentication and billing. </p><p><strong>Network effects.</strong> Some products get more valuable as more people use them. Shared databases, community contributions, aggregated insights. A skill is individual. There&apos;s no mechanism for one person&apos;s use to improve another person&apos;s experience, unless you <em>actually build</em> that mechanism.</p><p>*always is an interesting concept when using LLM&apos;s for most anything</p><h2 id="the-honest-assessment-of-impactpath">The honest assessment of ImpactPath</h2><p>So is ImpactPath just a skill? Would Death by Clawd give it a high mortality rating?</p><p>100%, yes. The core value, guiding someone through theory of change creation could absolutely be a skill. Someone with Claude Pro could get 80% of the value by loading an impact-framework skill and having the same conversation.</p><p>What the app adds:</p><ul><li>Saved progress (you can come back tomorrow)</li><li>Polished visual outputs (the diagram generation is fiddly)</li><li>A front door for people who don&apos;t use AI tools directly</li><li>Eventually: funder-specific templates, organisational accounts, portfolio views for funders</li></ul><p>Is that enough to justify the app? Honestly, I&apos;m don&apos;t think so. The saved progress is nice but not essential for a process most people complete in one sitting. The visual outputs could arguably be skill-generated. The front door matters if your users aren&apos;t already in AI-native workflows, which admittly many users still aren&apos;t yet.</p><p>The portfolio view for funders might be the thing that actually requires an app. That&apos;s aggregate data across multiple organisations, shared state, access control. There&apos;s probably collaborative infrastructure potential, but realising that means people investing in this as a concept.</p><h2 id="what-this-means-for-building">What this means for building</h2><p>I&apos;m not saying don&apos;t build apps. I&apos;m saying ask the question first.</p><p>When you have an idea, before you spin up a repo and start scaffolding authentication, ask: is this just a skill? Could I write a SKILL.md that captures the expertise, test it in Claude or Cursor, and see if it actually delivers the value?</p><p>If yes, maybe that&apos;s the product. Maybe that&apos;s enough. </p><p>And if it turns out you need more, scale, observability, services, security, network effects, you can build the infrastructure later. But you&apos;ll be building it on top of a skill that already works, that already encodes the expertise, that you&apos;ve already validated.</p><p>The skill becomes the foundation with the app becoming the distribution and interface layer on top.</p><h2 id="what-this-means-for-domain-expertise">What this means for domain expertise</h2><p>Here&apos;s where it gets interesting for organisations, not just builders.</p><p>Every organisation has embedded expertise. Previously, that expertise stayed trapped unless you built a product around it which requires funding, technical capability, ongoing maintenance, commercial ambition. Most organisations quite reasonably said &quot;that&apos;s not our core mission&quot; and the expertise stayed local.</p><p>Skills could potential change this.</p><p>What if packaging your expertise as a skill became a normal thing organisations did? Not as a product to sell, but as a contribution to a commons. The way you might write a guide or share a template, but machine-readable, executable, actually usable by anyone with an AI agent.</p><p>I wrote a piece called <a href="https://tomcw.xyz/building-blocks-for-just-in-time-software/">The Interface: Building Blocks for Just-in-Time Software</a> which takes this a new level, but perhaps skills are enough. </p><p>The value isn&apos;t in the delivery mechanism but in the knowledge about how to do something well. Skills make that knowledge portable.</p><p>I wonder if there is value in skills discovery, either automated or workshopped, understanding the various ways across an organisation that people encode expertise naturally in their everyday. Could these be skills that could be adopted across an organisation?</p><p>Not every idea needs to be an app. Not every workflow needs a SaaS. Not every piece of expertise needs infrastructure wrapped around it. Sometimes a SKILL.md file is the product. Sometimes the answer to &quot;are you just a .md file?&quot; is: yes, and that&apos;s fine.</p><p>So next time you&apos;re about to build something, ask yourself: is this just a skill?</p>]]></content:encoded></item><item><title><![CDATA[The Grant Application Is Dead. What Comes Next?]]></title><description><![CDATA[How federated protocols, local agents, and organisational self-sovereignty could replace the broken funding model.
Come down the rabbit hole with me...]]></description><link>https://tomcw.xyz/the-grant-application-is-dead-what-comes-next/</link><guid isPermaLink="false">69c3ab584e4f0e383144c916</guid><category><![CDATA[Data]]></category><category><![CDATA[ai]]></category><category><![CDATA[Open Infrastructure]]></category><category><![CDATA[Tech]]></category><category><![CDATA[Systems]]></category><dc:creator><![CDATA[Tom Watson]]></dc:creator><pubDate>Wed, 25 Mar 2026 17:24:50 GMT</pubDate><media:content url="https://tomcw.xyz/content/images/2026/03/grant-application-is-dead.pptx--1--1.png" medium="image"/><content:encoded><![CDATA[<img src="https://tomcw.xyz/content/images/2026/03/grant-application-is-dead.pptx--1--1.png" alt="The Grant Application Is Dead. What Comes Next?"><p>A foundation puts out a call. Organisations write applications. A panel reads them, scores them, funds some, rejects most. This process has been the backbone of grant-making for decades. It was designed for a world of information scarcity &#x2014; where funders needed a structured mechanism to surface what organisations do, what they need, and whether they&apos;re credible.</p><p>That world is gone.</p><p>Large language models have made it ridiculously easy to produce polished sounding grant applications. Funders are now drowning in submissions many of them &apos;well-written&apos;, &apos;well-structured&apos;, and indistinguishable from one another, a soup of pretty good. The response has been predictable: use LLMs on the other side too, to sift and score. We&apos;re building a system within a system,  machines writing applications for machines to read,  and the human decisions that funding is supposed to enable are getting further away, not closer.</p><figure class="kg-card kg-image-card"><img src="https://tomcw.xyz/content/images/2026/03/grant-application-is-dead.pptx.png" class="kg-image" alt="The Grant Application Is Dead. What Comes Next?" loading="lazy" width="960" height="540" srcset="https://tomcw.xyz/content/images/size/w600/2026/03/grant-application-is-dead.pptx.png 600w, https://tomcw.xyz/content/images/2026/03/grant-application-is-dead.pptx.png 960w" sizes="(min-width: 720px) 720px"></figure><p>This isn&apos;t a problem we can optimise our way out of. The application form was an approach and technology for filtering in a low-bandwidth information environment, where the capacity to do the thing meant you only did the application if it you were really committed to it, or at least reasonably committed. That environment no longer exists. We need to think about what comes after.</p><h2 id="the-problem-isnt-volume"><strong>The problem isn&apos;t volume</strong></h2><p>It&apos;s tempting to frame this as a scale problem - too many applications, not enough capacity. But in my opinion it&apos;s a deeper issue that is more structural. The traditional model is built on information asymmetry as a filtering mechanism. The form exists to make applicants <em>prove</em> they&apos;re worthy. LLMs broke that filter by making proof cheap.</p><p>The application as form approach jumbles several things that should be separate: <strong>identity</strong> (who is this organisation?), <strong>evidence</strong> (what have they done and learned?), <strong>intent </strong>(what do they want to do next?), and <strong>fit</strong> (does this align with what we fund?). <br>Bundling all of these into a single, pass/fail, one-shot document written to a deadline, shaped by word counts, and performed for an audience of assessors. Of course there are some who debundle these, into EOI&apos;s and multi stage approaches, but that was a solution to when filling in a form was a capacity issue.</p><h2 id="is-there-another-way"><strong>Is there another way?</strong></h2><p>What if, instead of organisations applying to funders, the relationship ran the other way?</p><p>There&apos;s a school of thought that says we should just do away with applications entirely and become fully relational in grant-making. No forms, no calls, no competitive rounds. Just relationships. Funders get to know organisations over time, understand their work, build trust, and fund on the basis of that relationship. Approaches like <a href="https://www.farmingthefuture.uk/?ref=tomcw.xyz" rel="noreferrer">Farming the Future</a> and <a href="https://regenerativefuturesfund.org.uk/the-fund?ref=tomcw.xyz" rel="noreferrer">Regenerative Futures</a> have moved in this direction, more connected, more trust-based, more human.</p><p>There&apos;s a lot to like about this. It removes the performative nature of the application form and centers the relationship. It lets funders understand context in a way that no written submission ever captures.</p><p>But can every foundation take this approach? And what are the potential risks if they do? Fully relational funding, without any balancing infrastructure, risks reverting to the boys&apos; club. Who gets the relationships? Who gets invited into the room? Predominantly, it&apos;s those with existing access to networks and that access is overwhelmingly a function of privilege. Geography, education, professional background, confidence in certain social settings. The organisations doing the most important work in the most underserved communities are often the furthest from funder networks. A purely relational model doesn&apos;t just fail to fix this. It can make it worse.</p><p>So the choice maybe isn&apos;t &quot;applications or relationships?&quot; but perhaps it&apos;s &quot;how do we build infrastructure that enables relationships to form on more equitable terms?&quot;</p><p>That&apos;s what leads to the idea of a <strong>self-sovereign organisational profile</strong>. Imagine a world where each organisation maintains a living, structured, machine-readable representation of itself. Not one written for any specific funder or call, but maintained for its own purposes. It covers the things funders always want to know: what the organisation does, who it serves, how it&apos;s governed, its financial health, its evidence of impact, what it&apos;s learning, its culture and ways of working. It&apos;s updated continuously, fed by the organisation&apos;s own systems, not composed under deadline pressure.</p><p>Funders, in this model, become discoverers rather than gatekeepers. They search across a landscape of organisations, exploring profiles, understanding context, and reaching out when they see alignment. The application form disappears and in its place: a structured, ongoing, mutual relationship between resource and purpose. The infrastructure makes organisations <em>discoverable</em> regardless of whether they already have a relationship with the funder. </p><p>This isn&apos;t entirely new thinking. Elements of it exist in open data initiatives, in the 360Giving standard, in community foundation practice. But the enabling technologies of federated protocols, structured data standards, LLM-assisted search and sense-making are now mature enough to make the full vision practical.</p><h2 id="an-organisational-knowledge-layer"><strong>An organisational knowledge layer</strong></h2><p>The core idea is a self-sovereign organisational profile: owned by the organisation, hosted on their terms, and selectively shared with funders and others who request access.</p><p>Think of it as an <strong><em>org.txt</em></strong> for the social sector but richer, living, and <strong><em>federated</em></strong>. Part narrative, part structured data. Markdown for the human-readable story. JSON or YAML for the machine-readable structure. Together, they form a profile that can be read by a person, queried by software, and understood by an LLM.</p><p>The profile covers several domains:</p><p><strong>Identity</strong> &#x2014; who the organisation is, its legal form, registration details, geography, scale, mission statement, and values. The stable facts that change slowly.</p><p><strong>Evidence</strong> &#x2014; what the organisation has done and what it has learned. Service delivery data, outcomes, case studies, evaluations. This is longitudinal, accumulating over time, telling a story about trajectory rather than presenting a snapshot.</p><p><strong>Governance</strong> &#x2014; how decisions are made, who&apos;s involved, what policies are in place. Board composition, safeguarding, financial controls. The things that signal organisational health and due diligence.</p><p><strong>Culture</strong> &#x2014; how the organisation works, what it values, how it treats people. This is harder to structure but matters enormously. It might include approaches to participation, equity commitments, learning practices.</p><p><strong>Ideas</strong> &#x2014; and this is where it gets interesting.</p><figure class="kg-card kg-image-card"><img src="https://tomcw.xyz/content/images/2026/03/grant-application-is-dead.pptx--1-.png" class="kg-image" alt="The Grant Application Is Dead. What Comes Next?" loading="lazy" width="960" height="540" srcset="https://tomcw.xyz/content/images/size/w600/2026/03/grant-application-is-dead.pptx--1-.png 600w, https://tomcw.xyz/content/images/2026/03/grant-application-is-dead.pptx--1-.png 960w" sizes="(min-width: 720px) 720px"></figure><h2 id="the-ideas-layer"><strong>The ideas layer</strong></h2><p>Right now, the grant model assumes organisations are reactive, sitting there waiting to be asked and then respond within the constraints of a funding call. But organisations are <em>always</em> thinking about what they&apos;d do with resource. They are the ones closest to communities, and have ideas sitting in people&apos;s heads, in strategy documents, in board away-day notes, in conversations with people. These ideas are often invisible to funders until someone writes them into an application.</p><p>What if ideas were first-class objects in the system?</p><p>An organisation publishes an idea &#x2014; structured, tagged, place or theme-based, loosely costed, linked to their evidence and track record. Not a full proposal, but a more like a seed. It sits there, discoverable, evolving. The organisation can refine it over time, link it to emerging evidence, connect it to ideas from other organisations working in the same place or theme.</p><p>Funders browse the landscape of ideas, not applications. They see clusters of intent forming , six organisations across two places all circling similar work. They can <strong><em>signal</em></strong> interest, which is itself visible and logged. They might approach a cluster and say: &quot;We&apos;d fund this. Let&apos;s talk.&quot; Or they might see an idea that&apos;s been developing for two years, gathering evidence and collaborators, and recognise that it&apos;s ready.</p><p>This changes the temporality of funding fundamentally. Instead of artificial deadlines and competitive rounds, funding becomes responsive to organisational readiness and ecosystem dynamics, shifting power. </p><p>This is sort of something I&apos;ve been exploring through<a href="https://openideas.uk/?ref=tomcw.xyz"> <u>OpenIdeas.uk</u></a> , the concept that ideas should be visible, connectable, and persistent, not locked inside application forms that only one funder ever sees.</p><h2 id="federation-no-platform-just-a-network"><strong>Federation: no platform, just a network</strong></h2><p>The worst version of this idea is a centralised platform, another portal where organisations create profiles, controlled by whoever runs it. Stop me if you&apos;ve heard this story before. It creates new gatekeepers, new lock-in, new dependencies.</p><p>The better version is federated. Each organisation hosting its own profile or has it hosted on their behalf, and the profiles communicate via an open protocol. No one owns the index. No single entity decides who&apos;s visible. The organisation <em>is</em> its own node in the network.</p><p>Two protocols are serious candidates: <a href="https://atproto.com/?ref=tomcw.xyz" rel="noreferrer">AT Protocol</a> (the technology behind Bluesky) and <a href="https://activitypub.rocks/?ref=tomcw.xyz" rel="noreferrer">ActivityPub</a> (the W3C standard behind Mastodon and the wider fediverse).</p><h3 id="at-protocol"><strong>AT Protocol</strong></h3><p>AT Protocol&apos;s strongest feature for this use case is account portability. An organisation&apos;s identity  based on a decentralised identifier (DID) travels with them. If a hosting provider disappears or becomes problematic, the organisation takes their data and moves. For organisations that have experienced <a href="https://tomcw.xyz/why-we-created-techfreedom-and-why-we-think-its-important/" rel="noreferrer">platform lock-in</a>, this is compelling. The DID also provides cryptographic verification of identity, which matters in a funding context where trust is essential.</p><p>The <a href="https://atproto.com/specs/lexicon?ref=tomcw.xyz" rel="noreferrer">lexicon system</a> is also relevant. AT Protocol lets you define custom record types meaning you could define an idea, an evidence-summary, an access-grant as formal lexicons, and they become first-class objects in the network. The <a href="https://docs.bsky.app/docs/advanced-guides/firehose?ref=tomcw.xyz" rel="noreferrer">firehose architecture</a> means funders could subscribe to a real-time stream of updates across the network as new ideas are published, evidence updated, profiles changed. The flow of information becomes continuous rather than periodic.</p><p>But there are limitations too. AT Protocol is still heavily shaped by Bluesky&apos;s social media needs. The infrastructure is resource-intensive to self-host which reintroduces some centralisation at the infrastructure layer and most critically for our purposes, AT Protocol assumes a public-by-default model. The tiered access control we might need, of public summaries, detailed profiles visible only to approved funders, full evidence bases available on request would need to be built as an additional layer rather than being native to the protocol. </p><h3 id="activitypub"><strong>ActivityPub</strong></h3><p><a href="https://en.wikipedia.org/wiki/ActivityPub?ref=tomcw.xyz" rel="noreferrer">ActivityPub</a> brings maturity and ecosystem breadth. The actor model maps naturally to organisations with each organisation being an actor that publishes activities: &quot;new idea published,&quot; &quot;evidence updated,&quot; &quot;access granted&quot; to followers. Funders follow organisations. ActivityPub handles private and semi-private content more naturally. You can address activities to specific actors or collections, which maps well to tiered access. The inbox/outbox model provides a built-in audit trail of interactions.</p><p>The downsides: identity is server-bound, if an organisation&apos;s server goes down, their identity goes with it. No native portability. </p><h3 id="a-pragmatic-decision-protocol-agnostic-with-a-bias"><strong>A pragmatic decision: protocol-agnostic, with a bias</strong></h3><p>Neither protocol is perfect as-is. AT Protocol offers better identity, portability, and structured data. ActivityPub offers better access control, maturity, and the pub/sub model we need.</p><p>So maybe the best path is to design the data model and vocabulary first, protocol-agnostic. Start with the data model. Get it right. Then implement federation beneath it.</p><p>The central artefact isn&apos;t a platform or a protocol but a <strong>local agent</strong> running on or for each organisation.</p><p>This is the organisation&apos;s node in the network. It pulls, it assembles, it federates. It does the work so people don&apos;t have to.</p><ul><li><strong>Pulls evidence from existing systems</strong> &#x2014; CRM data, service delivery records, financial information, impact metrics. Organisations don&apos;t maintain yet another system; the agent connects to what they already use via APIs, MCP servers, database connections, and document indexing.</li><li><strong>Maintains the living profile</strong> &#x2014; assembling structured data and narrative into the canonical representation. An LLM assists with summarisation, structuring, and keeping things current, but the organisation retains editorial control.</li><li><strong>Manages ideas</strong> &#x2014; providing a space to draft, refine, tag, and publish ideas, linked to evidence and to ideas from other organisations.</li><li><strong>Handles access control</strong> &#x2014; processing requests from funders, logging who sees what, managing tiered permissions. The organisation decides. The agent enforces.</li><li><strong>Speaks to the protocol</strong> &#x2014; federating the profile outward via ActivityPub, AT Protocol, or both. The agent is where the protocol-agnostic data model meets the specific federation mechanism.</li></ul><p>For some organisations, the agent runs on their own infrastructure. For others, it&apos;s a managed service hosted on their behalf but with their data remaining sovereign. The key is that the organisation owns the agent&apos;s output regardless of where it runs.</p><p>This architecture also means the agent can grow incrementally. An organisation might start with just a static profile generated from their website and Companies House data (something like what<a href="https://llmstxt.social/?ref=tomcw.xyz"> <u>llmstxt.social</u></a> already does). Over time, they connect their CRM, start publishing ideas, grant access to funders. Start small and build it up. </p><h2 id="what-this-enables"><strong>What this enables</strong></h2><p><strong>For organisations:</strong> maintain your own truth, once, and share it on your terms. Stop rewriting the same information for every funder. Make your ideas visible before anyone asks. See who&apos;s looking at your work. Connect your ideas to others doing complementary work.</p><p><strong>For funders:</strong> discover organisations doing relevant work, rather than waiting for them to find your call. See patterns and clusters across places and themes. Fund into ecosystems, not isolated bids. </p><p><strong>For the system:</strong> Make smaller organisations with strong practice more visible, not less. Create mutual transparency about who holds power and how it&apos;s exercised. Build a shared evidence layer that benefits everyone.</p><h2 id="some-caution-obviously"><strong>Some caution obviously</strong></h2><p>This isn&apos;t without hard problems. Some of them:</p><p><strong>Capacity.</strong> Even with low-friction tooling, some organisations won&apos;t be able to maintain a profile. The system needs to account for this </p><p><strong>Equity by design, not afterthought.</strong> If the system makes it easier for well-resourced organisations to maintain rich, polished profiles while smaller organisations struggle to keep theirs current, we&apos;ve just moved the privilege barrier from &quot;who writes the best application&quot; to &quot;who maintains the best profile.&quot; </p><p><strong>Governance of the standard.</strong> Who maintains the schema? How does it evolve? This needs to be stewarded by a body that represents both organisations and funders, and ideally by people who understand open standards development.</p><p><strong>Preventing surveillance.</strong> If funders can see everything, this could become a new <a href="https://en.wikipedia.org/wiki/Panopticon?ref=tomcw.xyz" rel="noreferrer">panopticon</a>. The access control layer and audit trail are essential, not optional. Organisations must be able to see exactly who has looked at what, and revoke access. The power relationship must be genuinely mutual.</p><p><strong>Interoperability with existing systems.</strong> The CRM landscape alone is fragmented Salesforce, Lamplight, Charitylog, Airtable, spreadsheets. The local agent needs robust MCP integrations for all of these. </p><p><strong>Funder adoption.</strong> Funders would need to move from designing calls and reading applications to searching, discovering, and entering dialogue. The system probably needs to coexist with traditional grant-making for a long transition period.</p><h2 id="building-on-what-exists"><strong>Building on what exists </strong></h2><p>This isn&apos;t starting from zero. But it&apos;s worth being honest about the limitations of what already exists, because they point toward what needs to be different.</p><p><strong>Open Referral UK and HSDS</strong> provide a standard for describing services what&apos;s available, where, for whom. But Open Referral is defined by <em>services</em>, not by <em>organisations and their ideas</em>. It assumes things are relatively static: a service exists, it has a location and eligibility criteria, you can look it up. That&apos;s useful for directories, but that&apos;s not how organisations or ideas are in reality. </p><p><strong>360Giving</strong> has demonstrated that open grant data is possible and valuable. But it describes funding <em>after the fact</em> who gave what to whom. It doesn&apos;t address the  question of how resource finds purpose in the first place.</p><p><strong>The Charity Commission</strong> is an interesting case. They&apos;re increasingly requiring (or planning to) organisations to report on &quot;<a href="https://www.charitysorp.org/?ref=tomcw.xyz" rel="noreferrer">impact</a>&quot; whatever that means  (this could be its own post entirely, because the gap between what regulators mean by impact and what organisations experience as meaningful change is vast). But the direction of travel is clear: more structured reporting, more data, more accountability. The use case driving this is regulatory. Not necessarily improving anything for organisations or the people they work with.</p><p>And so, if things are heading this way anyway shouldn&apos;t we try to do it better? Shouldn&apos;t we build the infrastructure so that the same structured data that satisfies a regulatory requirement <em>also</em> makes an organisation discoverable to funders, <em>also</em> connects their ideas to potential collaborators, <em>also</em> tells a richer story than a tick-box compliance exercise?</p><h2 id="what-comes-next"><strong>What comes next</strong></h2><p>This is an early exploration, not a finished concept, the writing sometimes is the thinking. If it were left to me maybe I be thinking about the following: define the schema properly, with input from organisations and funders. Build a reference agent, starting simple, with a static profile and manual updates, adding system integrations over time. Then test it with real relationships: a willing funder, a small cohort of organisations, a genuine alternative to a funding round. See what breaks.</p><p>Or, we can just keep going round and round ever chasing efficiencies and optimising a broken system. </p>]]></content:encoded></item><item><title><![CDATA[The Case for Loose Ends]]></title><description><![CDATA[Is our obsession with neat little bows on workshops and projects actually diminishing the impact they have? The long-term impact. The impact we hope they have beyond our involvement?]]></description><link>https://tomcw.xyz/the-case-for-loose-ends/</link><guid isPermaLink="false">69b98dd24e4f0e383144c80b</guid><category><![CDATA[Culture]]></category><category><![CDATA[Learning]]></category><category><![CDATA[Open Working]]></category><dc:creator><![CDATA[Tom Watson]]></dc:creator><pubDate>Wed, 18 Mar 2026 14:13:07 GMT</pubDate><media:content url="https://images.unsplash.com/photo-1644432304166-52a42d78a5a6?crop=entropy&amp;cs=tinysrgb&amp;fit=max&amp;fm=jpg&amp;ixid=M3wxMTc3M3wwfDF8c2VhcmNofDYzfHx0aHJlYWRzfGVufDB8fHx8MTc3Mzg0MzAzNnww&amp;ixlib=rb-4.1.0&amp;q=80&amp;w=2000" medium="image"/><content:encoded><![CDATA[<img src="https://images.unsplash.com/photo-1644432304166-52a42d78a5a6?crop=entropy&amp;cs=tinysrgb&amp;fit=max&amp;fm=jpg&amp;ixid=M3wxMTc3M3wwfDF8c2VhcmNofDYzfHx0aHJlYWRzfGVufDB8fHx8MTc3Mzg0MzAzNnww&amp;ixlib=rb-4.1.0&amp;q=80&amp;w=2000" alt="The Case for Loose Ends"><p>Is our obsession with neat little bows on workshops and projects actually diminishing the impact they have? The long-term impact. The impact we hope they have beyond our involvement?</p><p>I&apos;ve been designing and running a fair few workshops recently. Structure is important. How we start, how things flow, and how we end. I&apos;ve been thinking about learning and application of ideas, I&apos;ve been thinking about what happens in the room and more importantly what happens outside the room, beyond your control as a facilitator. And I&apos;ve been wondering whether in our efforts to show a workshop &apos;works&apos;, we close off deeper learning, and how sometimes we need to intentionally leave open threads and open questions. </p><h2 id="the-problem-with-resolution">The problem with resolution</h2><p>Most of what happens in a workshop or a session is only ever useful if people can take it back into their own context and make sense of it there. A good facilitator will create rooms and spaces that support the workshop. But however well designed, the room is artificial. The real work happens when someone is back at their desk on a Tuesday morning trying to figure out what any of it means for the decision they&apos;re actually facing.</p><p>If we resolve everything in the session with neat actions, clear conclusions, a satisfying arc, there is the potential we&apos;ve artificially done the sense-making for them. We&apos;ve removed the productive friction of trying to figure out what it means to me in my context outside the room. We&apos;ve made it easy to file the experience away. </p><p>But what if we leave a question hanging, one that is genuinely &amp; intentionally unresolved, not because we ran out of time but because we chose to? One that forces them to contextualise, to test an idea against their own reality, to keep thinking after the room has emptied.</p><p>Maybe the loose end isn&apos;t a failure of facilitation. In some cases maybe it&apos;s <em>the </em>mechanism.</p><h2 id="we-design-too-linearly">We design too linearly</h2><p>I think part of the problem is that we default to linear workshop design and I need to wrap up neatly. Input, discussion, activity, action plan. It feels productive. It maps well to a session plan. But sometimes it skips the slower, harder cognitive work &#x2014; the sitting with ambiguity, the divergent thinking, the wrestling with contradiction &#x2014; that actually changes how people see and operate.</p><p>I&apos;ve written about this in two frameworks &#x2014; the <a href="https://good-ship.co.uk/frameworks/?ref=tomcw.xyz">five modes of thinking</a> and the <a href="https://good-ship.co.uk/frameworks/modes-of-problem-solving.svg?ref=tomcw.xyz">modes of problem solving</a>. Both make the same core point: these aren&apos;t linear steps, they&apos;re modes you move between. But most workshop design treats them as a pipeline. We skip sensing and imagining, jump straight to designing and acting, and wonder why nothing sticks. We rush past the modes that require sitting with not-knowing because they&apos;re uncomfortable and hard to report on.</p><p>We&apos;re essentially <strong><em>optimising</em></strong> for what&apos;s legible in the room rather than what&apos;s <strong><em>transformative</em></strong> beyond it.</p><h2 id="honesty-about-complexity">Honesty about complexity</h2><p>There&apos;s a deeper issue too, particularly in systems change work. If we&apos;re genuinely working on complex, adaptive problems, the kind that don&apos;t have neat solutions, then wrapping a session up tidily is a kind of dishonesty. It implies that the problem can be bounded by a two-hour slot and a set of post-it notes.</p><p>The open question is the honest response to complexity. Saying &quot;we haven&apos;t resolved this, and that&apos;s OK, and here&apos;s why&quot; is more respectful of the work and of the people doing it than pretending we&apos;ve cracked something we haven&apos;t.</p><h2 id="intentional-not-accidental">Intentional, not accidental</h2><p>I do want to be clear though; this isn&apos;t an argument against structure, or for vague, meandering sessions where nothing happens, in fact I am a BIG advocate for structure and get a bit annoyed if I&apos;m in sessions where this doesn&apos;t happen! There&apos;s a world of difference between a loose end that exists because someone didn&apos;t plan properly and one that exists because a facilitator made a deliberate choice to leave space for ongoing sense-making.</p><p>The skill is knowing what to close and what to leave open. Naming it. Being honest with people that some things are meant to stay unresolved, that the discomfort of an open question is intentional and feeling comfortable with that. It&apos;s not always easy. People like certainty. </p><p>And to be clear, not every workshop or session needs this. Sometimes things can be neat, things do need resolving. The skill, and the bravery, is in knowing which one and designing for it. </p><h2 id="what-im-trying-to-do-differently">What I&apos;m trying to do differently</h2><p>In the organisational resilience work, and the TechFreedom sessions we are creating, I&apos;m experimenting with this. Designing sessions that explicitly don&apos;t resolve. Ending with questions rather than actions. Creating arcs that span weeks, not hours, so that the space between sessions becomes the real learning environment,  not dead time between the important bits.</p><p>It&apos;s harder to justify. It&apos;s harder to evidence. It looks less impressive on paper and it feels <em>uncomfortable</em>, uncertain. But I think it&apos;s closer to how change actually works &#x2014; slowly, messily, in the gaps between the things we can point to.</p><p>Maybe the best thing a workshop can do is send someone away with a question they can&apos;t stop thinking about, a niggle that they need to resolve in their time, in their context. Maybe that&apos;s not a loose end, but actually the whole point of it all. Maybe this is the thread worth following.</p>]]></content:encoded></item><item><title><![CDATA[What has the IMD ever done for us?]]></title><description><![CDATA[No-one is arguing that the Index of Multiple Deprivation is bad, but does it actually make a difference? 

A genuine question that's been bugging me for a while, and so, with a bit of apprehension, I wrote about why. ]]></description><link>https://tomcw.xyz/what-has-the-imd-ever-done-for-us/</link><guid isPermaLink="false">69b18ed54e4f0e383144c647</guid><category><![CDATA[Data]]></category><category><![CDATA[Neighbourhoods]]></category><category><![CDATA[policy]]></category><category><![CDATA[Questions]]></category><dc:creator><![CDATA[Tom Watson]]></dc:creator><pubDate>Wed, 11 Mar 2026 20:04:48 GMT</pubDate><content:encoded><![CDATA[<p>I&apos;m admittedly a little apprehensive about writing this blog, but it&apos;s a question that has been on my mind for a while. There&apos;s a couple of reasons for this apprehension. Firstly, it&apos;s pretty much a given that The Index of Multiple Deprivation (IMD) is a good thing with data, and I don&apos;t want to lose my good data card. Secondly, maybe the answer to my question is so blindingly obvious, but I&apos;m just not smart enough to see it. But this has been bugging me for a while, and so here I go...</p><p>I&apos;ve been using the Index of Multiple Deprivation for years. I&apos;ve pointed people to it. I&apos;ve built products with it. I&apos;ve laid it on maps. We all have right? You&apos;ve probably cited it in reports, used it to justify funding bids, and watched it shape commissioning decisions across the country.</p><p>And yet I&apos;ve always had this niggling question: has it actually made anything better?</p><p>This isn&apos;t a technical critique. I&apos;m not here to argue about the weighting of domains or the methodology behind the seven indicators. Smarter people than me have done that work. What I&apos;m asking is something more fundamental: has the existence of the IMD this tool we&apos;ve embedded so deeply into how policy gets made across the UK actually improved outcomes for the communities it describes?</p><h2 id="the-case-for">The case for</h2><p>Some would argue, yes, of course, stop asking ridiculous questions Tom. The IMD introduced a multidimensional way of understanding deprivation,  not just income, but health, education, employment, housing, crime, the living environment. It helped us think about intersectionality across different dimensions of need. It gave us a shared language and a common dataset that everyone from local councils to national government could point to.</p><p>And that&apos;s a good thing. Having a consistent, open dataset that covers England is genuinely useful. But has it made a <em>difference</em>?</p><h2 id="the-uncomfortable-evidence">The uncomfortable evidence</h2><p>In 2021, the Institute for Community Studies at The Young Foundation published something that should have landed like a bombshell. They found <a href="https://youngfoundation.b-cdn.net/wp-content/uploads/2021/07/ICS-WHY-DONT-THEY-ASK-US-compressed.pdf?x49395=&amp;ref=tomcw.xyz" rel="noreferrer">a 0% change in the relative economic advantage of the UK&apos;s most deprived neighbourhoods over 15 years, despite targeted investment of over &#xA3;20 billion across consecutive Labour, coalition, and Conservative governments.</a></p><p>Zero percent over fifteen years when we&apos;ve had the data use. </p><p>The most deprived places in 2004 were, by and large, the most deprived places in 2019. <a href="https://pmc.ncbi.nlm.nih.gov/articles/PMC9667433/?ref=tomcw.xyz" rel="noreferrer">Research looking at deprivation trajectories from 1971 to 2020</a> found : around 82% of areas in the most deprived decile in 2004 were still there in 2010, and by 2015-2019, nearly 88% of the most deprived areas stayed in that bottom decile.</p><p>So here&apos;s my question: if the IMD has been the primary thing guiding where billions of pounds of investment go, and the places it identifies as most deprived remain most deprived decade after decade, what exactly is it doing for us?</p><h2 id="a-static-snapshot-dressed-as-understanding">A static snapshot dressed as understanding</h2><p>The IMD is updated roughly every five years. It&apos;s built from administrative data, benefit claims, educational attainment, health records, crime statistics. It tells us about communities through the lens of what government systems already collect. It&apos;s a portrait drawn from the paperwork.</p><p>And it&apos;s a <em>relative</em> measure. It can tell you one area is more deprived than another, but not by how much. It can&apos;t tell you whether things are getting better in absolute terms. An area could see genuine improvements in the lives of its residents and still sit in the same decile because everywhere else improved too. There&apos;s also another problem, and this is a big one for me: The IMD describes areas, not people. </p><h3 id="false-certainty">False certainty?</h3><p>I sometimes wonder if the IMD gives us a false sense that we know what&apos;s happening and can control it? </p><p>Think about how it gets used in practice. A commissioner pulls up the IMD map or uses one of the many tools (more on this later) that list the areas they need to worry about.  They see the red areas, they write it into their strategy document, doing the right thing. There&apos;s a comfortable certainty to it: the data says <em>this</em>, so we do <em>that</em>. Is this just New Public Management wrapped up in maps?</p><p>But does the IMD actually tell us about what&apos;s happening in those communities? Does it tell us about the relationships between people? About the assets they hold, the community groups, the informal networks, the knowledge and capability that exists in every place? Does it capture what people care about, what they&apos;re trying to build, what keeps them up at night?</p><p>And if I&apos;m certain about anything, which at the moment is not a whole lot, it&apos;s that real change on a local level only comes from people who live there having real agency.</p><p>The IMD tells policymakers about communities. It doesn&apos;t enable communities to tell their own story. And there&apos;s a world of difference between those two things.</p><h2 id="when-certainty-stops-us-learning">When certainty stops us learning</h2><p>Here&apos;s what I think might be the real cost. When we have a dataset that feels authoritative and comprehensive, seven domains! thirty-nine indicators! every small area in England! we tend to stop asking questions. We stop being curious. We stop going to places and listening to people. Because we already <em>know</em>. The data told us.</p><p>And so we stop learning.</p><p>We wrap new programmes, new funding rounds, new policy initiatives in the same dataset, the same statutory data repackaged in a different way. <a href="https://www.neighbourhoodscommission.org.uk/pride-in-place-data-explorer/?ref=tomcw.xyz" rel="noreferrer">The Independent Commission on Neighbourhoods just released a new dashboard about Neighbourhoods</a>, it&apos;s nice, but guess what, it&apos;s the IMD with a couple of other indexs. </p><figure class="kg-card kg-image-card"><img src="https://tomcw.xyz/content/images/2026/03/Dexter-Idk-GIF.gif" class="kg-image" alt loading="lazy" width="450" height="450"></figure><p>New project, new programme management framework, different logo. But the same underlying assumption: that understanding deprivation means looking at the numbers, and that looking at the numbers means we know what to do. And this isn&apos;t even a knock on ICON, we&apos;ve all done it, and again, maybe I&apos;m just not smart enough to know that this really does make a difference. </p><h2 id="so-what-instead">So what instead?</h2><p>I&apos;m not arguing we should throw the IMD away. Data matters. Understanding patterns of deprivation at scale matters. But I am arguing we need to be honest about what it is and what it isn&apos;t.</p><p>The IMD is a backward-looking snapshot of administrative data. It&apos;s useful for understanding broad patterns. It&apos;s not a substitute for understanding communities, and it&apos;s certainly not a basis for designing interventions.</p><p>Perhaps what we should be doing is leaning into the uncertainty. Acknowledging that no dataset, however well-constructed, can capture the complexity of what&apos;s happening in a place. That the right response isn&apos;t to point at a map and say &quot;there do something about <em>that</em>&quot; but to go to that place and ask &quot;what&apos;s going on here? What do you need? What are you already doing?&quot; <a href="https://tomcw.xyz/maps-as-conversations/">Maps as conversations</a></p><h2 id="an-invitation">An invitation</h2><p>I&apos;m genuinely asking these questions, not just rhetorically. I&apos;ve used the IMD extensively and I suspect many reading this have too. So I&apos;d love to know:</p><p>Has it changed something for the better in your experience? Has it led to a decision that wouldn&apos;t have been made otherwise, one that actually improved things? Or has it become just something we do? A box we tick, a map we show, a certainty we lean on because the alternative,  admitting we don&apos;t really know is too uncomfortable?</p><p>I think there&apos;s a conversation to be had here. One about data and power, about who gets to define a place, and about whether the tools we&apos;ve built to understand deprivation might actually be getting in the way of doing something about it.</p>]]></content:encoded></item><item><title><![CDATA[Exploring conversational database building]]></title><description><![CDATA[Building databases through MCP servers with Airtable and Baserow and why open source alternatives make this stuff work better. Also giving a real example of the TechFreedom lens applied to two specific tools]]></description><link>https://tomcw.xyz/exploring-conversational-database-building/</link><guid isPermaLink="false">69b0232a4e4f0e383144c5cb</guid><category><![CDATA[ai]]></category><category><![CDATA[Data]]></category><category><![CDATA[Open Infrastructure]]></category><dc:creator><![CDATA[Tom Watson]]></dc:creator><pubDate>Tue, 10 Mar 2026 21:00:46 GMT</pubDate><media:content url="https://tomcw.xyz/content/images/2026/03/1000020424.png" medium="image"/><content:encoded><![CDATA[<img src="https://tomcw.xyz/content/images/2026/03/1000020424.png" alt="Exploring conversational database building"><p>Another in the series of blogs where I try to demystify things I&apos;ve been building and working on, and why they might matter beyond the technical details, but also diving into some of the TechFreedom lens. This time: connecting AI to databases, hitting walls with Airtable&apos;s API, and why open source makes more things possible.</p><p>I&apos;ve been spending a lot of time working within Airtable over the last week or so for a project.  I&apos;ve used Airtable for a good number of years and think it&apos;s genuinely good software, intuitive, user-friendly, powerful for a lot of use cases. But like everything, it&apos;s not without its limitations, like when you need to create large bases with complex data structures. Doing it all through the UI becomes a bit of an annoyingly slow process. Clicking through field after field, table after table, configuring options one at a time. Airtable lets you create bases from csv and their AI assistant is actually pretty good, but when things get beyond simple these methods don&apos;t really work.</p><p>As this was what I was actually doing, I decided to build something to solve the immediate problem. A JSON uploader that let me define a data structure in JSON and push it into Airtable via their API. Define your tables, fields, and relationships in a structured format, run the script, and the base is built. I&apos;m honestly surprised Airtable haven&apos;t done this themselves.</p><p>So far so good. But I wondered if I could go further obviously! The uploader was good, but a little inflexible. It worked perfectly for the exact JSON structure I&apos;d defined, but it was slightly limited to specific structures and so to make tweaks, change something, or handle a slightly different schema, I was back to either modifying the uploader or doing things manually. I could have built an adapter into the uploader (and still might) which allows this flexibility, but my mind went somewhere else first...</p><h2 id="enter-mcp-servers">Enter MCP servers</h2><p>This is where I started thinking about a different approach. Rather than building specific tools for specific tasks, what if I could create a persistent connection between an AI system and Airtable, one that could understand the base, work with it conversationally, and do whatever I needed, when I needed it?</p><p>If you&apos;re not familiar with MCP (Model Context Protocol), here&apos;s the short version: it&apos;s a standard way for AI tools to connect to external services. Think of it as a translator that sits between something like Claude or another AI assistant and a service like Airtable. The AI talks in natural language, the MCP server translates that into API calls, and suddenly you can say &quot;create a table called Organisations with fields for name, postcode, and a linked record to Services&quot; and it just... does it.</p><figure class="kg-card kg-image-card"><img src="https://tomcw.xyz/content/images/2026/03/mcp-how-it-works.png" class="kg-image" alt="Exploring conversational database building" loading="lazy" width="2000" height="1125" srcset="https://tomcw.xyz/content/images/size/w600/2026/03/mcp-how-it-works.png 600w, https://tomcw.xyz/content/images/size/w1000/2026/03/mcp-how-it-works.png 1000w, https://tomcw.xyz/content/images/size/w1600/2026/03/mcp-how-it-works.png 1600w, https://tomcw.xyz/content/images/2026/03/mcp-how-it-works.png 2400w" sizes="(min-width: 720px) 720px"></figure><p>Airtable already has an official MCP server. It lets you do certain things, search and analyse data, create records, update existing ones, but it didn&apos;t do ALL the things I wanted like actually build bases. So I took that as a starting point, extended it, and gave myself more options. I could now connect Claude or Mistral, or Gemini or a local model directly into Airtable and work with it properly, not just reading what was in a base, but creating tables, creating fields, adding records, restructuring things. Conversational database management, essentially.</p><p>And it was incredibly useful. For the kind of work I do building data structures for organisations, setting up CRM-like systems, creating directories, being able to describe what I need, provide JSON structures, markdown docs of definitions etc and have it built is a significant step up from clicking through forms.</p><h2 id="hitting-the-walls">Hitting the walls</h2><p>But then I started running into limitations. Not with the MCP server I&apos;d built, but with what Airtable&apos;s API actually allows you to do.</p><p>You can&apos;t create views via the API. You can&apos;t create interfaces. You can&apos;t create certain field types, formula fields being the big one, which are actually really fundamental to how most complex Airtable bases work. You can list and remove views, but you can&apos;t create them programmatically. </p><p>Now, these aren&apos;t technical limitations in the traditional sense. Airtable could expose these capabilities. They&apos;ve chosen not to. And when you think about why, it makes a kind of commercial sense. Every capability you can&apos;t automate is a capability that keeps you in their UI. Every limitation in the API is a reason you need to keep logging in, keep your team on paid accounts, keep building within their ecosystem rather than around it.</p><p>This is a strategic decision, not a technical one.</p><figure class="kg-card kg-image-card"><img src="https://tomcw.xyz/content/images/2026/03/api-comparisons.png" class="kg-image" alt="Exploring conversational database building" loading="lazy" width="960" height="540" srcset="https://tomcw.xyz/content/images/size/w600/2026/03/api-comparisons.png 600w, https://tomcw.xyz/content/images/2026/03/api-comparisons.png 960w" sizes="(min-width: 720px) 720px"></figure><h2 id="baserow-the-open-source-alternative">Baserow: the open source alternative</h2><p>This got me thinking about alternatives, specifically <a href="https://baserow.io/?ref=tomcw.xyz">Baserow</a>. It&apos;s an open source, EU built alternative to Airtable.</p><p>Baserow is API-first. That means every feature is designed to be an integration endpoint. When I extended the Baserow MCP server, I found I could do significantly more than with Airtable. The API is more complete because it&apos;s not designed to keep you inside a walled garden, it&apos;s designed to be useful. Creating views, managing fields including formulas, working with the full range of capabilities that the UI offers. The gap between what you can do in the interface and what you can do via the API is much, much smaller..</p><p>Now of course even saying all that, Baserow isn&apos;t perfect either. It doesn&apos;t have every feature Airtable has. The ecosystem is smaller, the polish isn&apos;t quite there in places. But the foundations are different in ways that potentially make it more useful for some organisations, it&apos;s all just about choosing what you value.</p><h2 id="the-techfreedom-lens">The TechFreedom lens</h2><p>This experience maps really well onto the kind of thinking we do in <a href="https://techfreedom.org.uk/?ref=tomcw.xyz">TechFreedom</a>, the programme Doug Belshaw and I are helping mission-driven organisations examine their technology dependencies. We look at technology choices through five lenses:</p><figure class="kg-card kg-image-card"><img src="https://tomcw.xyz/content/images/2026/03/techfreedom-five-lenses.png" class="kg-image" alt="Exploring conversational database building" loading="lazy" width="2000" height="1125" srcset="https://tomcw.xyz/content/images/size/w600/2026/03/techfreedom-five-lenses.png 600w, https://tomcw.xyz/content/images/size/w1000/2026/03/techfreedom-five-lenses.png 1000w, https://tomcw.xyz/content/images/size/w1600/2026/03/techfreedom-five-lenses.png 1600w, https://tomcw.xyz/content/images/2026/03/techfreedom-five-lenses.png 2400w" sizes="(min-width: 720px) 720px"></figure><p><strong>Jurisdiction.</strong> Where is your data? Who has legal authority over it? Airtable is US-based, data primarily in the US. Baserow is Dutch, data stored in Amsterdam.</p><p><strong>Business continuity.</strong> What happens if the vendor disappears or changes their pricing? With Baserow, the MIT licence means the software continues regardless. Your data is in PostgreSQL, not a proprietary format.</p><p><strong>Lock-in.</strong> How easy is it to leave? When your API won&apos;t let you recreate your views, formulas, or interfaces programmatically, migration is harder than it needs to be.</p><p><strong>Surveillance and data access.</strong> Who else can see your data, and under what legal frameworks? US-hosted and EU-hosted data sit under different rules.</p><p><strong>Cost exposure.</strong> Airtable&apos;s per-seat pricing scales quickly. Baserow starts free, paid from $10 per user per month. Self-hosting eliminates per-user costs entirely.</p><p>None of these lenses give you a simple answer. What they aim to do is give you a framework for thinking about what you&apos;re trading when you pick a tool.</p><p>When you&apos;re choosing tools, the question isn&apos;t just &quot;what can this do today?&quot; It should be &quot;what will this let me do tomorrow?&quot; Proprietary platforms set that ceiling based on their commercial interests. Open platforms set it based on what you&apos;re willing to build. For organisations watching their budgets and trying to stay resilient, I think that difference adds up.</p><hr><p>If you&apos;re thinking about any of this stuff database choices, AI integrations, technology dependencies or if you want to explore what TechFreedom could look like for your organisation, get in touch, find me on <a href="https://www.linkedin.com/in/tomcampbellwatson/?ref=tomcw.xyz">LinkedIn</a>, or sign up to the TechFreedom cohort!</p>]]></content:encoded></item><item><title><![CDATA[Why we created TechFreedom, and why we think it's important]]></title><description><![CDATA[Launching TechFreedom a cohort based programme to help you make choices about your tech. Why we created it, why we think it's important, what it's all about and an invite to join us!]]></description><link>https://tomcw.xyz/why-we-created-techfreedom-and-why-we-think-its-important/</link><guid isPermaLink="false">69a9ec244e4f0e383144c46f</guid><category><![CDATA[Tech]]></category><category><![CDATA[Strategy]]></category><category><![CDATA[Data]]></category><category><![CDATA[Open Infrastructure]]></category><dc:creator><![CDATA[Tom Watson]]></dc:creator><pubDate>Fri, 06 Mar 2026 09:16:48 GMT</pubDate><media:content url="https://tomcw.xyz/content/images/2026/03/Screenshot-2026-03-06-09.14.36-1.png" medium="image"/><content:encoded><![CDATA[<img src="https://tomcw.xyz/content/images/2026/03/Screenshot-2026-03-06-09.14.36-1.png" alt="Why we created TechFreedom, and why we think it&apos;s important"><p>
I&apos;ve spent years working with organisations who exist to do good in the world. Organisations whose missions are rooted in justice, community, care, and equity. And almost without exception, the digital infrastructure those missions run on belongs to companies who don&#x2019;t always share the same values.</p><h2 id="technology-is-a-culture-choice">Technology is a culture choice</h2><p>I wrote recently a short linkblog riffing on a great post from RebootDemocracy asking <a href="https://tomcw.xyz/linkblog/who-will-shape-ai-in-the-public-interest/"><u>who will shape AI in the public interest</u></a>. The post makes the case that governments have extraordinary power to shape how technology companies behave through procurement, and they&apos;re largely failing to use it.</p><p>What I liked most was how clearly it articulated something I&apos;ve argued for a long time: <strong>technology choices are cultural and political choices</strong>. Procurement isn&apos;t just a financial and risk management process, even though it is often treated as such. The tools you choose shape how your organisation works, what it values, and who holds power. Choose Microsoft and you&apos;re choosing a particular culture, one that in my opinion tends towards closure, risk aversion, and a narrowing of what feels possible. Choose alternatives and maybe you open up different ways of working, collaboration, different relationships with data, different assumptions about who gets to decide. Now maybe that IS how you want to work, or maybe you are willing to accept the tradeoff for other reasons, but did you ever make this choice?</p><p><strong>Technology is never neutral</strong>, and now with AI entering the picture, the cultural shaping gets even more pronounced. Microsoft probably leads to Copilot, maybe to OpenAI. Google leads down a different path, but one designed to keep you in that ecosystem. The way models respond, <a href="https://tomcw.xyz/people-as-code-collapse/"><u>what they surface and what they don&apos;t, starts to shape how people think</u></a>. Every choice is a cultural one, even when it&apos;s never described that way.</p><p>And it&apos;s not just about <em>what</em> you choose. It&apos;s about whether anyone in the room even realises there was a choice to be made in the first place</p><h2 id="lists-of-alternatives-are-nice-but-they-dont-shift-things">Lists of alternatives are nice, but they don&apos;t shift things</h2><p>About a year ago (around the time Elon was tearing out the heart of digital government in the USA) I shared a site listing <a href="https://tomcw.xyz/linkblog/european-alternatives/"><u>European alternatives</u></a> to US-based cloud services.&#xA0; It&apos;s a great resource,&#xA0;everything from hosting to domain registrars to VPNs and browsers. Whether you think the concern about US tech infrastructure is overblown or not, I said at the time there were three good reasons to look: over-reliance on a few companies is never good; there are genuinely excellent tools in there; and have you looked at your risk register lately?</p><p>The link between knowing and acting, between operation and strategic can be a bridge too far. A list of alternatives,&#xA0; however good, doesn&apos;t really change behaviour on its own. Maybe people look, nod, bookmark it, and then go back to the tools they know. Not because they&apos;re lazy or don&apos;t care, but because switching can feel hard, the risks feel abstract, and nobody has time to figure it all out alone.</p><p>That&apos;s the gap <a href="https://techfreedom.eu/?ref=tomcw.xyz"><u>TechFreedom</u></a> is trying to fill. Not another list, but a structured space to actually work through what your dependencies are, what risks they carry, and what you want to do about it.</p><h2 id="its-not-easy-and-were-not-pretending-it-is">It&apos;s not easy, and we&apos;re not pretending it is</h2><p>The ethics of big tech are not simple. Most organisations use the tools they use for good reasons, they work, they&apos;re familiar, and they&apos;re often subsidised or free** for mission-driven organisations. We&#x2019;re not interested in shaming anyone for using any particular technology, it&apos;s simply about making invisible dependencies visible so you can make deliberate choices, at your own pace.</p><p>I&apos;ll be honest about my own practice too. I try to use open source wherever I can.  This blog is on Ghost because it&apos;s open source. It was easier to be on substack honestly, but I made the choice, deliberately.  I build tools openly, share frameworks under Creative Commons, and default to open data. But I still use plenty of US-based infrastructure, because it&#x2019;s often the easiest or even the best technology for that purpose. The point of all this isn&apos;t purity. The point is being honest about the trade-offs and making them deliberately rather than by default.</p><p>When it comes to building things, especially with AI, I&apos;ve been quite vocal about designing for adaptability, using multiple AI platforms and API&apos;s deliberately, not because I think any one of them is utterly evil*, but because designing for multiple platforms builds resilience. </p><p>The <a href="https://techfreedom.eu/manifesto/?ref=tomcw.xyz"><u>TechFreedom manifesto</u></a> puts it plainly: no single company should have the power to stop an organisation working overnight. Essential services need clear exit paths, backups, and alternatives. If I only build on one provider&apos;s API and they change their pricing, their terms, or their politics, I&apos;m stuck. The same logic applies to any organisation&apos;s technology stack. <strong><em>Resilience means having options</em></strong>. It means knowing what your exit paths look like before you need them.</p><h2 id="we-built-techfreedom-on-european-infrastructure-it-was-harder">We built TechFreedom on European infrastructure. It <em>was</em> harder.</h2><p>When Doug and I set up techfreedom.eu, we made a deliberate choice to host on European infrastructure. The domain is European. The hosting is European. The tools we use for the programme itself are privacy-respecting and, where possible, open.</p><p>Honestly, it was just a bit harder to do. Not dramatically harder, but enough friction that I understand why most people don&apos;t bother. The defaults all pull you towards US platforms, they&apos;re often smoother, better documented, more integrated with everything else. Choosing differently takes a bit more effort, a bit more research, a few more decisions.</p><p>That friction is itself the problem. And it&apos;s one of the reasons we think the programme matters. If even people who do this for a living find it takes extra effort, imagine how it feels for an operations manager at a small organisation who just needs things to work.</p><h2 id="risk-registers-and-the-questions-nobody-asks">Risk registers and the questions nobody asks</h2><p>In my organisational resilience work, I&apos;ve spent years helping organisations think about how to anticipate, prepare, respond and adapt. Looking ahead, trying to give yourself options. What are your single points of failure? What&apos;s your plan B?</p><p>In my experience, most organisations have never asked these questions about their technology. They&apos;ve never mapped every tool they depend on and scored those tools against real risk lenses like <strong>jurisdiction, continuity and surveillance</strong>. Tech decisions are often made on technical proficiency and cost. They&apos;ve often never looked at their tech stack the way they&apos;d look at their finances or their safeguarding, as something that needs active governance.</p><h2 id="this-is-done-better-together">This is done better together</h2><p>We&#x2019;ve been quite intentional about TechFreedom being cohort-based. People working through the same process together over three sessions. You learn as much from hearing how someone else is navigating the same challenges as you do from any framework or facilitator. You spot blind spots in each other&apos;s thinking.&#xA0;</p><p>Practically it&apos;s three two-hour sessions, spaced a few weeks apart. In the first, you map your technology stack, all of them, including the ones you&apos;ve forgotten about. In the second, you score those dependencies against five risk lenses: j<strong>urisdiction, continuity, surveillance, lock-in, and cost exposure</strong>. In the third, you make some choices, priorities or a flexible roadmap, quick wins now, planned transitions over the next year, and strategic shifts over the longer term.</p><p>You leave with a clear picture of where you are, where your risks sit, and a practical plan for what to do about it. Plus a cohort of peers who are on the same journey, and people to keep you accountable.</p><p>If that feels like something you want to be a part of <a href="https://techfreedom.eu/?ref=tomcw.xyz" rel="noreferrer">read more here</a> or leave your email in the form below and we&#x2019;ll be in touch!</p><figure class="kg-card kg-bookmark-card"><a class="kg-bookmark-container" href="https://driftforms.app/f/techfreedom-sign-up-BRp1Gi?ref=tomcw.xyz"><div class="kg-bookmark-content"><div class="kg-bookmark-title">TechFreedom sign up</div><div class="kg-bookmark-description"></div><div class="kg-bookmark-metadata"><img class="kg-bookmark-icon" src="https://driftforms.app/icon.svg?icon.0ab92f7d.svg" alt="Why we created TechFreedom, and why we think it&apos;s important"><span class="kg-bookmark-author">Drift</span></div></div><div class="kg-bookmark-thumbnail"><img src="https://driftforms.app/f/techfreedom-sign-up-BRp1Gi/opengraph-image?3edb98d276c3076b" alt="Why we created TechFreedom, and why we think it&apos;s important"></div></a></figure><p>*I mean some might genuinely be evil</p><p>**free to a point and even then those corporations can change their terms quickly without much warning (yes I&apos;m looking at you Microsoft)</p>]]></content:encoded></item><item><title><![CDATA[Bridging the Data Gap: A Semantic Translation Layer for UK Poverty Data]]></title><description><![CDATA[Ponderings on how to bridge the data gap. Exploring questions, semantic layers, TRE's and feedback loops.]]></description><link>https://tomcw.xyz/bridging-the-data-gap-a-semantic-translation-layer-for-uk-poverty-data/</link><guid isPermaLink="false">6988ba27ac80fd04af71d1c0</guid><category><![CDATA[Data]]></category><category><![CDATA[Learning]]></category><category><![CDATA[Questions]]></category><dc:creator><![CDATA[Tom Watson]]></dc:creator><pubDate>Sun, 08 Feb 2026 16:43:05 GMT</pubDate><media:content url="https://tomcw.xyz/content/images/2026/02/data-bridge-slides.pptx--4-.png" medium="image"/><content:encoded><![CDATA[<img src="https://tomcw.xyz/content/images/2026/02/data-bridge-slides.pptx--4-.png" alt="Bridging the Data Gap: A Semantic Translation Layer for UK Poverty Data"><p>Last week I attended a workshop with RSS and JRF looking at Data Gaps in understanding poverty. I had a half (or quarter) baked concept, but I was so overly tired that I couldn&apos;t really form a well thought out way of explaining it. So I went down the rabbit hole after a bit of sleep.</p><hr><p>We talk a lot about data gaps in the UK,&#xA0; the idea that we don&apos;t have enough data to understand the issues we care about. But I think often, the problem isn&apos;t that the data doesn&apos;t exist, it&apos;s more often the case that the people who need it don&apos;t know it exists, can&apos;t find it, or can&apos;t tell whether it answers their question. Like many things, the gap isn&apos;t in the data, it&apos;s in the connection between data and need.</p><p>This is particularly acute for administrative data, the data collected by government departments as a byproduct of delivering services. Benefits claims, tax records, school meal eligibility, homelessness applications, GP registrations. This data is extraordinarily rich, but as I said, it&#x2019;s primary use isn&#x2019;t research, or understanding, it&#x2019;s for delivery services. And that means the way it is described and structured is for that purpose. It&apos;s also, for the most part, extraordinarily hard to discover and understand if you&apos;re not already embedded in the system that produced it.&#xA0;</p><p>But government departments aren&#x2019;t the only ones delivering services in this area. Local government and the VCSE are heavily involved in both delivery and trying to understand if our approaches are working</p><figure class="kg-card kg-image-card"><img src="https://tomcw.xyz/content/images/2026/02/data-bridge-slides.pptx--4--1.png" class="kg-image" alt="Bridging the Data Gap: A Semantic Translation Layer for UK Poverty Data" loading="lazy" width="960" height="540" srcset="https://tomcw.xyz/content/images/size/w600/2026/02/data-bridge-slides.pptx--4--1.png 600w, https://tomcw.xyz/content/images/2026/02/data-bridge-slides.pptx--4--1.png 960w" sizes="(min-width: 720px) 720px"></figure><h3 id="the-three-layers-of-disconnect">The three layers of disconnect</h3><p>When a VCSE organisation is trying to evidence fuel poverty for a funding bid, they&apos;re not going to search for &quot;sub-national fuel poverty statistics LSOA-level DESNZ (department for energy security and net zero for those following along at home).&quot; They&apos;re going to think: &quot;how many people in my area can&apos;t afford to heat their homes?&quot; And they&apos;re going to hit a wall, not because the data doesn&apos;t exist, but because nothing bridges their question to the answer.</p><p>The disconnect operates at three levels.&#xA0;</p><ul><li>First, existence: people don&apos;t know a dataset is out there at all.&#xA0;</li><li>Second, relevance: even if they find something, they can&apos;t tell whether it answers their specific question.&#xA0;</li><li>Third, usability: even if it&apos;s relevant, the format, geography, or timeliness doesn&apos;t work for them.</li></ul><p>Current data portals and catalogues - <a href="https://www.data.gov.uk/?ref=tomcw.xyz" rel="noreferrer">data.gov.uk</a>, <a href="https://stat-xplore.dwp.gov.uk/webapi/jsf/login.xhtml?ref=tomcw.xyz" rel="noreferrer">Stat-Xplore</a>, <a href="https://fingertips.phe.org.uk/?ref=tomcw.xyz" rel="noreferrer">Fingertips</a>, the ONS website mostly address only the first level, and they do it using keyword search against technical metadata. The metadata describes the data in the language of the producer, not the language of the user.</p><h3 id="what-a-bridge-would-look-like">What a bridge would look like</h3><figure class="kg-card kg-image-card"><img src="https://tomcw.xyz/content/images/2026/02/data-bridge-slides.pptx--6--1.png" class="kg-image" alt="Bridging the Data Gap: A Semantic Translation Layer for UK Poverty Data" loading="lazy" width="960" height="540" srcset="https://tomcw.xyz/content/images/size/w600/2026/02/data-bridge-slides.pptx--6--1.png 600w, https://tomcw.xyz/content/images/2026/02/data-bridge-slides.pptx--6--1.png 960w" sizes="(min-width: 720px) 720px"></figure><p>The idea is simple in concept: build a <strong>semantic translation layer</strong> that sits between user intent and data metadata. Instead of requiring users to learn the language of the data, the system learns the language of users&apos; needs.</p><p>&quot;Oh, you&apos;re trying to understand child poverty in your area? You&apos;d want DWP&apos;s children in low-income families data for the baseline, DfE&apos;s free school meals data gives you a school-level proxy, and you could triangulate with HMRC&apos;s child benefit data,&#xA0; but be aware that FSM eligibility criteria changed with Universal Credit, so the time series has a break.&quot;</p><p>That response does three things: </p><ul><li>it translates from the user&apos;s framing to relevant datasets, </li><li>it suggests combinations across departmental boundaries, </li><li>and it carries contextual caveats about data quality. </li></ul><p>These are the functions the bridge needs to perform.</p><h3 id="how-it-could-work-technically">How it could work technically</h3><p>The core technical approach uses embeddings, vector representations that capture semantic meaning,&#xA0; to match user questions to enriched dataset descriptions. Two things get embedded into the same vector space:</p><p><strong>User intents</strong> - the <strong>questions*</strong> people are trying to answer, expressed in their own language. &quot;I want to understand health inequalities in Gateshead.&quot; &quot;I need to evidence demand for youth services.&quot; &quot;I&apos;m writing a bid about digital exclusion among older people.&quot;</p><p><strong>Enriched metadata</strong> - not just the raw catalogue entry, but descriptions of what each dataset can tell you, what questions it&apos;s been used to answer before, what it measures and what it misses, and how it relates to other datasets.</p><p>*stop me if you&#x2019;ve heard me say this before&#x2026;</p><p>When a user poses a question, the system embeds it, finds the nearest metadata vectors, and uses an LLM to generate a plain-language explanation of why each dataset might be relevant including caveats and suggestions for complementary sources.</p><p>The crucial addition is a <strong>feedback loop</strong>. Every time a user says &quot;yes, that dataset was useful&quot; or &quot;no, that wasn&apos;t what I needed,&quot; the system learns. Over time, it builds a rich understanding of which data helps answer which kinds of questions,&#xA0; knowledge that currently exists only in the heads of a few specialist analysts.</p><p>Now this use of questions is something I&apos;ve talked a lot about, and previously we at Data For Action explored the idea of a <a href="https://open.substack.com/pub/dataforaction/p/introducing-our-newest-prototype?r=ed0wj&amp;utm_campaign=post&amp;utm_medium=web&amp;ref=tomcw.xyz" rel="noreferrer"><strong>Question Bank </strong></a>which attempted to surface the most important questions people were trying to answer with data. And we built a prototype which did some of this semantic meaning, grouping similar questions automatically. Where I think this idea builds upon that is by focusing on the link between questions and metadata and perhaps proposing an <strong>enhanced semantic metadata standard</strong>. </p><h3 id="starting-with-poverty-data">Starting with poverty data</h3><figure class="kg-card kg-image-card"><img src="https://tomcw.xyz/content/images/2026/02/data-bridge-slides.pptx--2-.png" class="kg-image" alt="Bridging the Data Gap: A Semantic Translation Layer for UK Poverty Data" loading="lazy" width="960" height="540" srcset="https://tomcw.xyz/content/images/size/w600/2026/02/data-bridge-slides.pptx--2-.png 600w, https://tomcw.xyz/content/images/2026/02/data-bridge-slides.pptx--2-.png 960w" sizes="(min-width: 720px) 720px"></figure><p>So I recognise this would require a fair bit of work across every government dataset. But as the workshop, and JRF&#x2019;s focus is on Poverty, let&#x2019;s explore what this might mean. Obviously the thing about poverty is that, well, it&#x2019;s inherently intersectional. Understanding poverty in a place requires data from DWP (benefits and income), HMRC (tax credits and earnings), DLUHC (housing costs and deprivation), DfE (free school meals as a proxy), NHS (health inequalities), ONS (census and labour market), and local authorities (council tax support, local welfare). No single department owns the picture.</p><p>This intersectionality forces the system to work across boundaries from day one,&#xA0; which is where both the challenge and the potential value lies.</p><h3 id="a-phased-approach">A phased approach</h3><p>Rather than trying to build the full system immediately, a staged rollout builds value and trust progressively.</p><p><strong>Phase 1: Internal discovery within a single department</strong> - Start within DWP. Help their analysts discover relevant datasets from other teams within their own department. This proves the enriched metadata and embedding approach, delivers immediate value, and requires no cross-departmental agreements. The key question to answer: does this help people find data they didn&apos;t know existed within their own organisation?</p><p><strong>Phase 2: Cross-departmental metadata sharing</strong> -&#xA0; Extend to share enriched metadata, not the underlying data, across departments. Start with DWP and HMRC as the highest-value pairing for poverty analysis. This requires metadata sharing agreements, not data sharing agreements, which is a significantly lighter governance ask. Policy teams working on intersectional questions gain visibility into what evidence exists across government.</p><p><strong>Phase 3: External access</strong> -&#xA0; Open the discovery and translation layer to researchers, local authorities, and VCSE organisations. This is where the feedback loop becomes most powerful, because a much wider and more diverse set of use-cases and questions flows back into the system.</p><p>Possibly the most interesting long-term implication is understanding and mapping the demand that emerges from the feedback loop. Over time, the system reveals patterns: &quot;Lots of organisations in the North East are trying to answer questions about X, but no good administrative data source exists.&quot; Or: &quot;This dataset is frequently matched to questions but users consistently report it isn&apos;t granular enough.&quot;</p><p>That&apos;s a powerful feedback signal to data producers. It makes the evidence-based case for what data should be published, at what granularity, with what frequency, and with what metadata. It closes the loop between data supply and data demand in a way that doesn&apos;t currently exist in the UK data landscape.</p><h3 id="beyond-discovery-rethinking-research-access">Beyond discovery: rethinking research access</h3><p>Everything described so far concerns published data, the aggregates, tables, and statistics that government departments make available through their websites and platforms. But some of the most important administrative data never makes it into published outputs. Record-level benefits data, linked tax and earnings records, individual-level health and education data. This is the data that could answer the most pressing questions about poverty, but accessing it currently requires navigating one of the most restrictive research infrastructure models in the world.</p><p>In the UK, detailed administrative data is primarily accessible through Trusted Research Environments - the ONS Secure Research Service, HMRC Datalab, DWP data sharing arrangements, and NHS Digital&apos;s Data Access Request Service. These often require a fully specified research proposal detailing your questions, methodology, required datasets and variables, and expected outputs. The process typically takes months. You need institutional affiliation and accreditation. The data never leaves the secure environment.</p><p>This model rests on a fundamental assumption: you must know exactly what you&apos;re looking for before you can look. You cannot just explore, poke at the edges. You cannot say &quot;I think the relationship between housing benefit claims and GP registration patterns might reveal something about how people fall through gaps between services, but I need to see the data to know.&quot; You arrive with a finished question or you don&apos;t arrive at all.</p><p>The result is a system optimised for a narrow type of user,&#xA0; experienced quantitative researchers at accredited institutions who already know the data landscape well enough to name the specific datasets and variables they need. Everyone else is excluded: local authority analysts, VCSE organisations, policy teams working across departmental boundaries, and researchers at earlier stages of inquiry who haven&apos;t yet refined their hypotheses.</p><figure class="kg-card kg-image-card"><img src="https://tomcw.xyz/content/images/2026/02/data-bridge-slides.pptx--7-.png" class="kg-image" alt="Bridging the Data Gap: A Semantic Translation Layer for UK Poverty Data" loading="lazy" width="960" height="540" srcset="https://tomcw.xyz/content/images/size/w600/2026/02/data-bridge-slides.pptx--7-.png 600w, https://tomcw.xyz/content/images/2026/02/data-bridge-slides.pptx--7-.png 960w" sizes="(min-width: 720px) 720px"></figure><p>The Data Bridge concept has the potential to shift this in three ways.</p><p><strong>Better-specified access requests</strong> Even within the existing TRE model, the bridge improves outcomes. A researcher who uses the translation layer to discover that a specific combination of DWP and HMRC variables would answer their question arrives at the application process with a stronger, more precisely scoped request. The bridge has done the preliminary investigation that currently requires either insider knowledge or months of desk research. This alone could significantly reduce the friction and failure rate of TRE applications.</p><p><strong>Aggregated demand as evidence for new outputs</strong> - This is where it gets genuinely interesting. Currently, if fifty organisations across the country all want to understand the relationship between in-work poverty and housing costs in their area, each one separately discovers (or fails to discover) that the relevant data exists, separately works out that they&apos;d need linked DWP and DLUHC data, and most give up because TRE access is beyond their reach. The bridge makes this demand pattern visible and quantifiable: &quot;There were 200 queries this quarter attempting to understand in-work poverty and housing costs at local authority level. No published dataset serves this need. The underlying data requires record-level access through separate TRE applications to two departments.&quot;</p><p>This creates an evidence base for creating a new standard output,&#xA0; a pre-linked, appropriately aggregated, routinely published dataset designed to serve demonstrated demand without requiring individual record-level access. The feedback loop builds the business case for data producers to invest in new publications by showing them exactly what users need and how often.</p><p><strong>A permissive middle tier: guided aggregation</strong> -&#xA0; The most ambitious implication is the possibility of a new access model sitting between open published tables and full TRE access. Call it <strong>guided aggregation</strong>. A user specifies their question through the bridge. The system identifies the relevant underlying datasets and variables. Rather than granting the user access to records, it runs pre-approved analytical queries against the data within the secure environment and returns aggregated, disclosure-controlled results tailored to the user&apos;s actual question.</p><p>The user never sees a record. The analysis is constrained to prevent re-identification. Statistical disclosure control is applied automatically. But the output is responsive to the specific question being asked, rather than being a one-size-fits-all published table that may not cut the data in the way the user needs.</p><p>This isn&apos;t unprecedented in concept. The <a href="https://www.abs.gov.au/statistics/microdata-tablebuilder/tablebuilder?ref=tomcw.xyz"><u>Australian Bureau of Statistics&apos; TableBuilder </u></a>operates on a similar principle, users construct custom tabulations from microdata, with automated confidentialisation. But combining this with a semantic translation layer that helps users formulate their questions and identify the right data sources would be a significant step beyond what currently exists.</p><p><strong>From research-ready to question-ready</strong> - The UK has invested substantially in making administrative data research-ready through TREs, the ADR UK programme, and the growing network of secure research infrastructure. This investment has been valuable. But research-ready is not the same as question-ready. <strong>Research-ready</strong> means the data is clean, documented, and available in a secure environment for accredited researchers with pre-approved proposals. <strong>Question-ready</strong> means the data is accessible at appropriate levels of aggregation and disclosure control to the full range of people who have legitimate questions it could help answer.</p><p>The shift matters because the people closest to the problems that poverty data should illuminate, frontline charities, community organisations, local authority officers, the people being described by the data are almost entirely excluded from the current access model. They have questions. The data has answers. The bridge isn&apos;t just about connecting the two but about building an infrastructure that treats their questions as legitimate grounds for access, not just the questions that arrive wrapped in a formal research proposal from a Russell Group university.</p><h3 id="what-this-isnt">What this isn&apos;t</h3><p>It&apos;s worth being clear about boundaries. This doesn&apos;t replace TREs or the safeguards they provide, record-level data access for complex research will always need secure environments and governance. It doesn&apos;t remove the need for analytical skills, statistical literacy, or human judgement about data quality. And the guided aggregation concept would need significant work on automated disclosure control and governance frameworks before it could operate at scale.</p><p>What it does is challenge the assumption that the only legitimate interaction with administrative data is a fully pre-specified research project. It suggests that<strong> discovery,</strong> <strong>exploration</strong>, and <strong>question-driven analysis</strong> are valid modes of engagement with public data,&#xA0; and that building infrastructure to support them would unlock value that the current model leaves on the table.</p><p>More practically, it takes the knowledge that currently lives in the heads of a small number of specialist intermediaries,&#xA0; the people who know what data exists, where to find it, what it can and can&apos;t tell you, and how different sources relate and makes that knowledge systematically accessible. And it takes the questions that thousands of organisations ask in isolation and makes them collectively visible, creating the demand signal that data producers need to justify investment in better, more accessible outputs.</p><hr><p>It&apos;s also likely that this isn&apos;t the solution, or it&apos;s simply unworkable. What I heard in the workshop was people who think about this a lot more than me, with a lot more knowledge than me. And so, who really am I to think I&apos;ve got the answer? Just someone who thinks questions are important, and that the biggest barrier to data isn&apos;t really data at all. <br></p>]]></content:encoded></item><item><title><![CDATA[The Interface: Building Blocks for Just-in-Time Software]]></title><description><![CDATA[The second of my 8 ideas to explore in 2026 in AI and technology. In this one I'm exploring just in time software, like lego blocks we pull together when we need them. ]]></description><link>https://tomcw.xyz/building-blocks-for-just-in-time-software/</link><guid isPermaLink="false">693c95252174e13db933bebc</guid><category><![CDATA[ai]]></category><category><![CDATA[Tech]]></category><category><![CDATA[Strategy]]></category><dc:creator><![CDATA[Tom Watson]]></dc:creator><pubDate>Sun, 01 Feb 2026 21:33:02 GMT</pubDate><media:content url="https://images.unsplash.com/photo-1575470522418-b88b692b8084?crop=entropy&amp;cs=tinysrgb&amp;fit=max&amp;fm=jpg&amp;ixid=M3wxMTc3M3wwfDF8c2VhcmNofDd8fGxlZ28lMjB8ZW58MHx8fHwxNzY5OTgxNDAwfDA&amp;ixlib=rb-4.1.0&amp;q=80&amp;w=2000" medium="image"/><content:encoded><![CDATA[<h1 id></h1><img src="https://images.unsplash.com/photo-1575470522418-b88b692b8084?crop=entropy&amp;cs=tinysrgb&amp;fit=max&amp;fm=jpg&amp;ixid=M3wxMTc3M3wwfDF8c2VhcmNofDd8fGxlZ28lMjB8ZW58MHx8fHwxNzY5OTgxNDAwfDA&amp;ixlib=rb-4.1.0&amp;q=80&amp;w=2000" alt="The Interface: Building Blocks for Just-in-Time Software"><p>The second of 8 ideas I set out in <a href="https://tomcw.xyz/predictions-for-2026-or-what-im-actually-thinking-about/" rel="noreferrer">this blog at the end of 2025</a></p><p>If I was going to build a startup in 2026 (and I might) here&apos;s what I would be looking at: not generative dashboards, but generative components that let people (and especially the social sector) build exactly what it needs, when it needs it.</p><h2 id="the-2023-prediction">The 2023 Prediction </h2><p>Back in 2023 (ish), I thought we&apos;d soon stop building static dashboards and instead just talk to our data. Ask questions, get visualisations on the fly. &quot;Show me referrals by area last month.&quot; Boom, chart appears.</p><p>And we&apos;re sort of there. You can absolutely do this now. Claude and GPT can generate charts from data. Ask a question, get a dashboard.</p><p>Except it doesn&apos;t really work. Not because the AI can&apos;t generate the interface. But because our data isn&apos;t ready for that really. Our data is still messy, siloed, inconsistently structured, and locked in systems that can&apos;t expose it cleanly or easily. Try spinning a just in time dashboard while going in a death loop of microsoft admin privileges.  </p><p>But as we move into 2026 I have begun wondering again about just in time interfaces beyond just dashboards, but as service components ready to be deployed.</p><h2 id="what-if-we-went-component-level-instead">What If We Went Component-Level Instead?</h2><p>What if we generated the building blocks that organisations could assemble into exactly the software they need, right now, for this specific task?</p><p>The social sector doesn&apos;t need fewer tools. It needs the ability to create specific tools when specific needs arise, without requiring &#xA3;100k custom development projects or forcing everyone onto generic one-size-fits-none platforms.</p><p>The startup opportunity is building a component library and assembly infrastructure specifically for just in time software and services. </p><h2 id="why-components-not-complete-applications">Why Components, Not Complete Applications</h2><p>Every organisation I work with has the same conversation:</p><p>&quot;We need software that does [incredibly specific thing relevant to our specific community, in our specific context, integrated with our specific other systems].&quot;</p><p>The options are:</p><ol><li>Pay a fortune for custom development</li><li>Use a generic platform and compromise on 70% of what you actually need</li><li>Glue together five different SaaS tools and spend all your time managing integrations</li><li>Build it yourself and hope</li></ol><p>What if there was option 5:</p><ol start="5"><li>Assemble it from standardised, interoperable components built by organisations who&apos;ve already solved similar problems</li></ol><p>Not &quot;here&apos;s a complete CRM,&quot; but &quot;here&apos;s the component for intake forms, the component for case tracking, the component for outcome measurement, the component for referral coordination&quot; and you assemble them into the specific tool your organisation needs.</p><h2 id="make-use-of-the-expertise-we-have">Make use of the expertise we have.</h2><p>This is where it gets interesting. The social sector has massive expertise embedded in specific contexts.</p><ul><li>A foodbank network that&apos;s solved complex inventory tracking in a crisis response context</li><li>A youth service that&apos;s built brilliant engagement tools for young people</li><li>A mental health charity that&apos;s developed sophisticated consent management for sensitive data</li></ul><p>Right now, that expertise stays trapped in their specific systems unless they decide to launch it into a sass product. </p><p>What if each of these organisations could extract the component they&apos;ve built and make it available for others to use in different contexts?</p><p>The foodbank&apos;s inventory component becomes useful for disaster response, community fridges, school uniform banks. The consent management component becomes useful across health, education, criminal justice sectors.</p><h2 id="the-startup-opportunity">The Startup Opportunity</h2><p><strong>The Component Library for Social Purpose</strong>: A registry of pre-built, open-source components designed specifically for social sector needs. Not generic business tools, but components that understand:</p><ul><li>Complex consent requirements</li><li>Safeguarding protocols</li><li>Outcome measurement frameworks</li><li>Partnership coordination</li><li>Resource scarcity</li><li>Community engagement</li></ul><p><strong>The Assembly Interface</strong>: An AI-powered tool that helps organisations describe what they need and assembles relevant components into a working prototype. &quot;We need to track community garden plots, coordinate volunteer shifts, and report outcomes to our funder&quot; &#x2192; the system suggests relevant components and shows how they&apos;d fit together.</p><p><strong>The Contribution Protocol</strong>: A way for organisations to package their specific solutions as reusable components. You&apos;ve built something brilliant? Extract the core functionality, generalise it just enough to be reusable, contribute it back to the commons.</p><p><strong>The Adaptation Layer</strong>: Because no component will be perfect for every context, an AI layer that helps adapt components to specific needs. The foodbank inventory component needs to work for hygiene products instead of food? The AI helps adapt the logic while maintaining the core patterns.</p><h2 id="what-this-looks-like-in-practice">What This Looks Like in Practice</h2><p>A small advice charity serving refugees needs:</p><ul><li>Intake forms in multiple languages</li><li>Case tracking with complex family relationships</li><li>Document management with privacy controls</li><li>Referral coordination with other services</li><li>Outcome measurement for funder reporting</li></ul><p>Traditional approach:</p><ul><li>Spend &#xA3;50k on custom development, or</li><li>Use a generic CRM and hack it into something sort-of-usable, or</li><li>Keep using spreadsheets and/or clunky databases</li></ul><p>Component approach:</p><ol><li>AI helps them describe their needs in plain language</li><li>System suggests relevant components:<ul><li>Multilingual intake forms (from a migrant support network)</li><li>Family relationship tracking (from a children&apos;s charity)</li><li>Secure document storage (from a domestic violence service)</li><li>Referral coordination (from the network protocol infrastructure)</li><li>Impact measurement (from a sector-wide outcomes framework)</li></ul></li><li>Organisation assembles these into a working prototype in hours, not months</li><li>AI helps adapt components to their specific context</li><li>They test with real users, refine, deploy</li><li>Six months later, they&apos;ve built a novel component for supporting refugees through housing applications and they contribute it back for others to use</li></ol><h2 id="why-this-might-actually-work-now">Why this (might) actually work now</h2><p>Three things make this viable now that weren&apos;t five years ago:</p><p><strong>1. AI can do the assembly and adaptation</strong>: The hard part isn&apos;t building components (developers can do that). It&apos;s helping non-technical people describe what they need and assemble components intelligently. LLMs are really good at this translation layer. Yes we will need to think carefully about the context layer, and constraining LLM&apos;s enough to make it work, but we have the tools to do it, and people, in both technical, data and service roles who could do this.</p><p><strong>2. APIs and microservices are mature</strong>: The technical patterns for building interoperable components are well established. We know how to do this now.</p><h2 id="could-this-be-the-way-to-solve-siloed-data">Could this be the way to solve siloed data?</h2><p>If components have standardised data structures, organisations gradually converge toward cleaner, more interoperable data. Not through mandated standards (which are really hard to make work in multi organisational settings), but through practical utility. &quot;This component needs data structured this way&quot; &#x2192; &quot;okay, let&apos;s structure it that way because this component is really useful.&quot; It&apos;s standards by stealth. </p><p>Bottom-up standardisation through shared components, not top-down mandates that everyone ignores.</p><h2 id="what-might-this-look-like">What might this look like</h2><p>In five years:</p><ul><li>A new charity can spin up working systems in days, not months</li><li>Innovations built by one organisation are available sector-wide</li><li>Small organisations have access to sophisticated tools previously only available to large ones</li><li>The social sector has a thriving commons of shared components</li><li>Organisations spend less time on systems and more time on mission</li><li>Data gradually becomes more interoperable through practical use, not mandates</li></ul><p>The infrastructure that makes this possible:</p><ul><li>Open source component library with social sector primitives</li><li>AI-powered assembly and adaptation tools</li><li>Clear protocols for contributing components back to the commons</li><li>Quality standards and peer review for component reliability</li><li>Governance by and for the sector, not by vendors</li></ul><h2 id="but-why-should-we-bother">But why should we bother?</h2><p>We are drowning in software debt and expensive systems that don&apos;t quite fit. We have technical debt we can&apos;t afford to fix and innovation locked in proprietary platforms.</p><p>We need infrastructure that lets the sector build what it needs, share what works, and adapt quickly when needs change.</p><p>Not another platform. The infrastructure for rapid, context-specific assembly of exactly what each organisation needs.</p><p>The AI can generate the interfaces. That&apos;s the easy part.</p><p>The hard part is building the component library, the contribution protocols, the quality standards, and the governance structures. Components, not platforms. Commons, not vendors. Just-in-time assembly, not one-size-fits-none.</p><p>That&apos;s the interface layer the sector actually needs.</p>]]></content:encoded></item><item><title><![CDATA[Building LLMs.txt for the social sector]]></title><description><![CDATA[Another blog on the ins and outs of something I built, what it's all about, and why it might be a useful tool and/or concept. A llms.txt tool designed for specific use cases. ]]></description><link>https://tomcw.xyz/building-llms-txt-for-the-social-sector/</link><guid isPermaLink="false">697c7e66ac80fd04af71d00f</guid><category><![CDATA[ai]]></category><category><![CDATA[Learning]]></category><category><![CDATA[Tech]]></category><category><![CDATA[tools]]></category><dc:creator><![CDATA[Tom Watson]]></dc:creator><pubDate>Fri, 30 Jan 2026 15:50:29 GMT</pubDate><media:content url="https://tomcw.xyz/content/images/2026/01/slide-2-how-it-works-1.png" medium="image"/><content:encoded><![CDATA[<img src="https://tomcw.xyz/content/images/2026/01/slide-2-how-it-works-1.png" alt="Building LLMs.txt for the social sector"><p>Or as I call it &quot;AI find my organisation and be accurate please&quot;</p><p>This is another in a series of blogs exploring things I&apos;ve built, lifting the lid on both the technical and conceptual ideas behind them. I hope it helps us in the social purpose sector think about technology &amp; AI as more than chat interfaces, and that we perhaps should be looking at things on a broader level.</p><p>This time I want to talk about infrastructure and boring tiny tools. Nope, no flashy chat here I&apos;m afraid.&#xA0; Instead, I want to explore something much more foundational: how do we make the work of social purpose organisations visible and accessible to AI systems in the first place?</p><h3 id="what-is-llmstxt"><strong>What is llms.txt?</strong></h3><figure class="kg-card kg-image-card"><img src="https://tomcw.xyz/content/images/2026/01/data-src-image-ce4f9ee4-9a98-432a-9b19-a62faf44eac6.png" class="kg-image" alt="Building LLMs.txt for the social sector" loading="lazy" width="800" height="500" srcset="https://tomcw.xyz/content/images/size/w600/2026/01/data-src-image-ce4f9ee4-9a98-432a-9b19-a62faf44eac6.png 600w, https://tomcw.xyz/content/images/2026/01/data-src-image-ce4f9ee4-9a98-432a-9b19-a62faf44eac6.png 800w" sizes="(min-width: 720px) 720px"></figure><p>Before we get into what I&apos;ve built, let&apos;s talk about the problem I&apos;m trying to solve.</p><p>AI systems - the large language models (LLMs) that power everything from ChatGPT to Claude to various coding assistants - increasingly need to understand websites. Whether it&apos;s answering questions about an organisation, helping someone find support services, or assisting a funder in understanding who&apos;s working in a particular space, these systems need to get information from somewhere.</p><p>The problem is that websites are designed for humans, not machines. They&apos;re full of navigation menus, JavaScript, cookie banners, and complex layouts. Converting a modern website into something an AI can actually understand and use is surprisingly hard. And even when you do extract the text, context windows (the amount of information an AI can process at once) are limited - and although they are increasing, you can&apos;t just feed it an entire website and hope for the best.</p><p>This is where llms.txt comes in. It&apos;s a proposed specification - created by<a href="https://www.answer.ai/posts/2024-09-03-llmstxt.html?ref=tomcw.xyz"><u> Jeremy Howard at Answer.AI</u></a> - for a simple markdown file that sits at `/llms.txt` on your website and provides a curated, AI-friendly summary of who you are and what you do. Think of it as a &quot;robots.txt for AI&quot; - but instead of telling crawlers what not to index, it tells AI systems what they actually need to know about you.</p><p>A standard llms.txt file has a specific structure: a title, a short description, some context, and then links to more detailed information organised into sections. It&apos;s human-readable (it&apos;s just<a href="https://en.wikipedia.org/wiki/Markdown?ref=tomcw.xyz"><u> markdown</u></a>) but also machine-parseable, which means tools can work with it programmatically.</p><h3 id="why-does-the-social-sector-need-this"><strong>Why does the social sector need this?</strong></h3><p>So we&apos;ve already<a href="https://www.cazenovecapital.com/en-gb/uk/charity/insights/how-ai-is-changing-search-and-what-it-means-for-google-chatgpt-and-the-open-web/?ref=tomcw.xyz"><u> seen the data, showing that traffic from traditional sources (search engines) is decreasing</u></a>, as people use LLMs as search, and so organisations without good AI-readable content will become increasingly invisible. If someone asks an AI assistant &quot;what charities support refugees in Newcastle?&quot; and your organisation doesn&apos;t show up because the AI couldn&apos;t understand your website properly, you&apos;ve lost an opportunity to help someone.</p><p>But there&apos;s a bigger picture too. If the social sector wants to build towards a future where data flows more freely - where funders can more easily understand the landscape, where collaboration happens because people can actually find each other, where we stop reinventing wheels because we didn&apos;t know the wheel already existed, then we need to consider things like this.</p><p>llms.txt is a small piece of infrastructure that supports this vision. It&apos;s a way for organisations to represent themselves clearly and consistently to AI systems, in a format that&apos;s standardised enough to be useful but flexible enough to capture the nuance of what social purpose organisations actually do.</p><h3 id="what-ive-built-and-why-its-different"><strong>What I&apos;ve built (and why it&apos;s different?)</strong></h3><p>I built the tool to generate llms.txt files automatically. It&apos;s called<a href="https://llmstxt.social/?ref=tomcw.xyz"><u> <strong>llmstxt-social</strong></u></a> and it exists in two forms: an open-source command line tool that anyone can run themselves, and a web-based service for those who&apos;d rather not deal with the technical bits.</p><figure class="kg-card kg-image-card"><img src="https://tomcw.xyz/content/images/2026/01/data-src-image-741b18b8-bf77-421a-ac7c-c150055be6e2.png" class="kg-image" alt="Building LLMs.txt for the social sector" loading="lazy" width="800" height="500" srcset="https://tomcw.xyz/content/images/size/w600/2026/01/data-src-image-741b18b8-bf77-421a-ac7c-c150055be6e2.png 600w, https://tomcw.xyz/content/images/2026/01/data-src-image-741b18b8-bf77-421a-ac7c-c150055be6e2.png 800w" sizes="(min-width: 720px) 720px"></figure><p>So, if you hit google(<a href="https://www.ecosia.org/)?ref=tomcw.xyz"><u>other search engines are available</u></a><u> </u>) and type in llms.txt generator, there are many. So why build another one?&#xA0; Well, the other generators are just based on the original llms.txt specification which is quite general - it was designed with technical documentation in mind. Our version adapts this for social purpose organisations with some important additions.</p><p><strong>Templates for different organisation types. </strong>A charity has different information needs than a funder, which has different needs than a local authority. We have four templates - charity, funder, public sector, and startup - each with sections that make sense for that type of organisation.</p><p>For instance, our charity template includes:</p><ul><li><strong>For Funders section</strong>: registration number, geography, themes, beneficiaries - exactly the information a grant-maker needs to assess fit</li><li><strong>For AI System </strong>section: explicit guidance on how AI should represent the organisation, including caveats and things to avoid</li><li><strong>Data enrichment</strong>- We don&apos;t just scrape the website - we pull in data from external sources to provide verified context:<ul><li><strong>Charity Commission data</strong> (official registration, financial information, charitable objects)</li><li><strong>360Giving data for funders </strong>(grants history, typical award sizes, geographic distribution)</li></ul></li><li>This means the llms.txt file contains information that might not even be on the website, but is crucial for understanding the organisation properly.</li><li><strong>Quality assessment. </strong>We don&apos;t just generate a file and leave you to it. The tool includes a comprehensive assessment system that scores completeness, checks for missing sections, and provides actionable recommendations for improvement.</li></ul><p><strong>How it actually works</strong></p><p>Let me break down what happens when you run the tool:</p><p><strong>Step 1: Crawling</strong></p><p>We start by crawling the website. This sounds simple but involves a fair bit of care:</p><ul><li>Respecting robots.txt (we&apos;re good AI citizens)</li><li>Using sitemap.xml if available (many sites have one but don&apos;t realise it)</li><li>Falling back to link discovery if not</li><li>Rate limiting to avoid hammering servers</li><li>Optionally using Playwright for JavaScript-heavy sites that don&apos;t render properly otherwise</li></ul><p>We typically limit this to around 30 pages - enough to get a good picture, not so many that we&apos;re processing irrelevant content.</p><p><strong>Step 2: Content extraction</strong></p><p>Each page gets processed to extract just the actual content:</p><ul><li>Removing navigation, headers, footers, scripts, styles</li><li>Classifying the page type (about, services, contact, projects, etc.)</li><li>Extracting contact information</li><li>Finding charity numbers</li></ul><p>This classification is important - it means we can group related pages together sensibly in the final output.</p><p><strong>Step 3: Data enrichment</strong></p><p>If we&apos;ve found a charity number (or one was provided), we call the Charity Commission API to get official data:</p><ul><li>Registered name and status</li><li>Registration date</li><li>Latest financial information (income and expenditure)</li><li>Charitable objects (the formal statement of what the charity does)</li><li>Trustee information</li><li>Contact details</li></ul><p>For funders, we can also pull 360Giving data showing their actual grants history - incredibly useful for applicants trying to understand what a funder really supports versus what their website says (or doesn&apos;t say) they support.</p><p><strong>Step 4: AI analysis</strong></p><p>Now comes the LLM bit. We feed the extracted content and enrichment data to Claude and ask it to:</p><ul><li>Write a concise mission summary</li><li>Generate clear descriptions for each page</li><li>Categorise services and programmes</li><li>Identify target beneficiaries</li><li>Create guidance for how AI systems should represent the organisation</li></ul><p>This is where the &apos;magic&apos; happens, but it&apos;s also where careful prompting matters. We use structured outputs (schema validation) to ensure the AI returns data in a consistent, predictable format rather than getting creative with the structure.</p><p><strong>Step 5: Generation</strong></p><p>Finally, we assemble all this into the llms.txt file using the appropriate template. The output follows the spec strictly - H1 title, blockquote description, H2 sections with markdown lists of links and descriptions.</p><p><strong>Step 6: Assessment (optional)</strong></p><p>If you want, we&apos;ll then assess the generated file against what we know about the organisation:</p><ul><li>Is it complete? Are expected sections present and populated?</li><li>Is it accurate? Does it align with the enrichment data?</li><li>Is it useful? Will an AI actually be able to use this effectively?</li></ul><p>We score out of 100 and provide specific recommendations for improvement.</p><p><strong>Provider agnosticism (again)</strong></p><p>Just like with<a href="https://tomcw.xyz/building-open-recommendations/"><u> Open Recommendations</u></a>, I&apos;ve built this to be AI provider agnostic. The tool currently defaults to Claude, but the architecture supports switching to other providers. Given how rapidly the AI landscape is changing, this flexibility is essential.</p><p>Different models are better at different things. The extraction and classification tasks could potentially run on smaller, cheaper models. The summary generation benefits from more capable models. Having the flexibility to mix and match - or to switch entirely if pricing or availability changes - is crucial for sustainability.</p><p><strong>The subscription model: monitoring for change</strong></p><p>Here&apos;s something that took me a while to figure out: llms.txt files need to stay current. Organisations change - new services launch, old ones close, contact details update. A stale llms.txt file is arguably worse than no file at all.</p><p>So i&apos;ve built a subscription tier that monitors your website and automatically regenerates your llms.txt file when things change significantly. It:</p><ul><li>Periodically recrawls your site</li><li>Compares new content against the existing file</li><li>Regenerates if meaningful changes are detected</li><li>Tracks change history so you can see what evolved</li></ul><p>This costs &#xA3;9/month - priced to be accessible to small organisations while covering the actual costs of running the infrastructure (hosting, API calls, monitoring).</p><p><strong>Open source vs paid service</strong></p><p>I want to be clear about the model here because I think it matters. <strong><em>No-one funds me to do this stuff.&#xA0;</em></strong></p><p><strong>The core tool is open source (MIT licensed).</strong> You can clone the repository, run it on your own machine, generate as many llms.txt files as you want, and never pay us a penny. If you&apos;re technically capable and have an Anthropic API key, there&apos;s nothing stopping you from doing everything yourself.</p><p><strong>The web service exists for convenience and for features that need ongoing infrastructure </strong>- like the monitoring subscription. The free tier gives you 10 generations per day (basic output). The paid one-off option (&#xA3;7) includes full assessment and enrichment. The subscription (&#xA3;9/month) adds monitoring and automatic updates.</p><p>The paid tiers are priced to cover costs - server hosting, API calls to Claude and the Charity Commission, database storage, monitoring infrastructure - plus a small margin to make this sustainable. We&apos;re not trying to extract maximum value; we&apos;re trying to make this genuinely accessible while keeping the lights on.</p><p><strong>What I&apos;ve learned and what I&apos;m trying to do with this stuff</strong></p><p>The things I&apos;m building and experimenting with at the moment are a quite intentional pushing at the edges and provocations for thinking about AI and us. Things I think (maybe):</p><p><strong>Infrastructure matters more than features:</strong> The social sector loves talking about AI for efficiency, AI for programme delivery, maybe for analysis.&#xA0; That&apos;s all important. But if our organisations aren&apos;t visible and correctly represented to AI systems in the first place, none of that matters. We need to invest in the boring plumbing as much as the shiny features.</p><p><strong>Standards are valuable even when imperfect: </strong>llms.txt is still a proposal. It might evolve, it might get superseded. But having <strong><em>something</em></strong> standardised to build on is immensely valuable. Perfect is the enemy of good, especially in infrastructure, minimum viable standards etc...</p><p><strong>The 80/20 split is real:</strong>&#xA0; No, I&apos;m not talking about my running training. As with Open Recommendations, about 80% of this project is not the AI part. It&apos;s the crawling logic, the validation, the error handling, the user interface, the subscription management, the deployment configuration. The AI is the exciting bit that makes it possible, but the engineering around it is what makes it usable.</p><p><strong>Open source plus commercial can work?:</strong> Can it? This is a test. Now of course there are many good examples of this actually working at scale, but what about small scale. I&apos;m a bit nervous/curious about this model - will people resent the paid tiers?&#xA0; Would it feel like bait-and-switch? So far, no, but it&apos;s early days and I&apos;ve managed to cover 2 months infrastructure costs, will that keep up? Being transparent about what&apos;s free and what costs money, and being honest about why things cost money, will hopefully work. But probably it all comes down to value, and if people don&apos;t value the tool, the model doesn&apos;t really matter.</p><h3 id="how-to-get-started"><strong>How to get started</strong></h3><p>Anyway, if you&apos;ve made it this far, maybe you want to try it out</p><p><strong>For the technical folks:</strong> Clone the repo, install the dependencies, set up your Anthropic API key, and run `llmstxt generate https://your-website.org.uk`. Full instructions in the README.</p><p><strong>For everyone else:</strong> Visit<a href="https://llmstxt.social/?ref=tomcw.xyz"><u> https://llmstxt.social,</u></a> paste in your URL, and click generate. The free tier will give you a basic llms.txt file. If you want the full assessment and enrichment, there&apos;s a one-off payment option. If you want ongoing monitoring, there&apos;s a subscription.</p><p>Either way, you&apos;ll end up with a file you can add to your website at `/llms.txt`. If you&apos;re using a CMS that makes this awkward, you can also link to it from your homepage or put it in a place you can reference.</p><h3 id="why-this-matters"><strong>Why this matters</strong></h3><p>I&apos;ll be honest - this isn&apos;t glamorous work. No one&apos;s going to get excited about infrastructure files for AI systems. There are no pretty dashboards, no impressive demos, no &quot;look what AI generated&quot; moments to share on social media.</p><p>But I genuinely believe this kind of foundational work is essential. As AI becomes more embedded in how people find and understand organisations, we need the social sector to show up properly. Not just to be visible, but to be represented accurately - with the nuance and context that our work deserves.</p><p>llms.txt is maybe one small piece of that puzzle. It&apos;s a way for organisations to take control of their AI representation rather than leaving it to chance. And by building tools that make it easy to create and maintain these files, I hope we&apos;re lowering the barrier enough that this becomes normal infrastructure for the sector.</p><p>And ff you have questions, thoughts, or want to try something like this and/or collaborate, give me a shout on <strong>tom@good-ship.co.uk </strong>or on the<a href="https://www.linkedin.com/in/tomcampbellwatson/?ref=tomcw.xyz"><u> linky</u></a><u>&#xA0;</u></p>]]></content:encoded></item><item><title><![CDATA[The Garden of Ideas]]></title><description><![CDATA[The Garden Ideas, a simple facilitation tool for individuals and groups.  Seeds, soil, seasons, pruning and composting as a way of making intentional choices]]></description><link>https://tomcw.xyz/the-garden-of-ideas/</link><guid isPermaLink="false">697b3fc8ac80fd04af71cfb6</guid><category><![CDATA[Strategy]]></category><dc:creator><![CDATA[Tom Watson]]></dc:creator><pubDate>Thu, 29 Jan 2026 12:05:28 GMT</pubDate><media:content url="https://tomcw.xyz/content/images/2026/01/garden-of-ideas--1-.jpg" medium="image"/><content:encoded><![CDATA[<blockquote>Ideas are easy...[insert one of the many quotes about execution/implementation/teams and stuff here]</blockquote><img src="https://tomcw.xyz/content/images/2026/01/garden-of-ideas--1-.jpg" alt="The Garden of Ideas"><p>I dunno. Are ideas easy? Maybe. Good ideas probably less so. But there is some truth that some of what stops ideas becoming more than that is choices and decisions. Often it&apos;s the choice to not do something, so that another idea can grow. And to grow an idea we often need to make decisions about when and how. </p><p>There are many tools and frameworks that can help you with these things, <strong>stop/start/continue</strong>, <strong>now/next/future</strong>,<strong> matrix</strong> of all types. So why another one Tom? I dunno, I guess I just kind of like gardens and metaphors.</p><p>So I put together The Garden Ideas, a simple facilitation tool for individuals and groups. </p><h3 id="first-the-metaphor">First, the metaphor.</h3><p>A garden isn&apos;t a holding space where things wait indefinitely. It&apos;s a living system that forces honest choices. Seeds are planted with intention (mostly though there always surprises or is that just me?). Some sprout quickly (hopefully). Others need time underground before they&apos;re ready. And crucially, not everything gets planted, a gardener makes choices about what the space can actually sustain and hopefully how the plants interact. Pollinators and fruit, complementary plants that sustain the soil etc.</p><p>Framing our ideas as seeds, our soil as conditions, our pruning or composting as choices seems to land with people. And in a world where our to-do list never shrinks, this approach can feel both kinder and more decisive. </p><p><strong>Composting feels kinder than deleting.</strong> When we decide an idea isn&apos;t right for this garden, this time, we&apos;re not throwing it away. We&apos;re returning it to the soil. It might nourish something else later. This matters when you&apos;re staring at a to-do list that&apos;s become a guilt list.</p><p><strong>Banking seeds is strategic, not procrastination.</strong> A seed bank isn&apos;t where things get forgotten, it&apos;s where they&apos;re preserved for the right conditions. This gives us permission to say &quot;not now&quot; without saying &quot;never.&quot;</p><p><strong>Seasons create natural timing conversations.</strong> Instead of debating whether an idea is good or bad, high priority or low, we can ask: is this the right season? Sometimes an idea needs conditions that don&apos;t exist yet. Sometimes we need to wait for capacity, funding, or the right people.</p><p><strong>Gardens have limits.</strong> A to-do list can grow forever. A garden can only hold so much. This constraint is a feature, it forces the prioritisation conversation that infinite lists let us avoid.</p><figure class="kg-card kg-image-card kg-card-hascaption"><img src="https://tomcw.xyz/content/images/2026/01/garden-of-ideas.jpg" class="kg-image" alt="The Garden of Ideas" loading="lazy" width="960" height="540" srcset="https://tomcw.xyz/content/images/size/w600/2026/01/garden-of-ideas.jpg 600w, https://tomcw.xyz/content/images/2026/01/garden-of-ideas.jpg 960w" sizes="(min-width: 720px) 720px"><figcaption><span style="white-space: pre-wrap;">The Garden elements - seeds, soil &amp; conditions, Seasons, Pruning and composting</span></figcaption></figure><h2 id="how-the-garden-works">How the garden works</h2><p>The framework scales from personal use to group facilitation.</p><p><strong>For individuals</strong>, it&apos;s straightforward: capture your seeds (ideas, tasks, possibilities), then sort them. What are you planting now&#x2014;actually committing time and energy to this season? What goes in the seed bank for when conditions change? What do you need to compost, finally letting go of things that have been sitting on lists for months?</p><figure class="kg-card kg-image-card"><img src="https://tomcw.xyz/content/images/2026/01/garden-of-ideas--2-.jpg" class="kg-image" alt="The Garden of Ideas" loading="lazy" width="960" height="540" srcset="https://tomcw.xyz/content/images/size/w600/2026/01/garden-of-ideas--2-.jpg 600w, https://tomcw.xyz/content/images/2026/01/garden-of-ideas--2-.jpg 960w" sizes="(min-width: 720px) 720px"></figure><p><strong>For groups</strong>, the framework adds a layer. Each person maintains their own garden throughout a session, capturing seeds as they emerge. A shared seed bank catches ideas that surface when the group can&apos;t explore them, and intentionally, you schedule when to return to it. At the close, people bring seeds from individual gardens into a collective space, deciding together what to plant, bank, or compost.</p><h2 id="the-sorting-conversation">The sorting conversation</h2><p>This is where the metaphor earns its keep. For each idea, ask:</p><ul><li><strong>Plant now?</strong> What does this need to grow? What resources, time, skills, people? Who will tend it?</li><li><strong>Seed bank?</strong> When will conditions be right? What needs to change first? When do we revisit?</li><li><strong>Compost?</strong> What would this free up if we let it go? What space does it create?</li></ul><p>The language gives permission to make real choices. &quot;Composting&quot; doesn&apos;t feel like failure, it feels like making space for what can actually flourish. And the seed bank offers a middle path between doing everything now and abandoning ideas entirely.</p><figure class="kg-card kg-image-card"><img src="https://tomcw.xyz/content/images/2026/01/garden-of-ideas--3-.jpg" class="kg-image" alt="The Garden of Ideas" loading="lazy" width="960" height="540" srcset="https://tomcw.xyz/content/images/size/w600/2026/01/garden-of-ideas--3-.jpg 600w, https://tomcw.xyz/content/images/2026/01/garden-of-ideas--3-.jpg 960w" sizes="(min-width: 720px) 720px"></figure><h2 id="a-word-on-pruning">A word on pruning</h2><p>Some facilitators struggle with groups who want to plant everything, or individuals who can&apos;t let anything go. The garden metaphor helps here too: a garden can only hold so much. Trying to grow everything means nothing gets the attention it needs.</p><p>Pruning isn&apos;t cruel. It&apos;s care (talking to myself while pruning the lavender....)</p><h2 id="using-the-toolkit">Using the toolkit</h2><p>I&apos;ve put together a simple toolkit with printable worksheets - an individual garden sheet, a seed bank poster, and a collective garden template. It&apos;s designed for multi-session facilitation but works just as well for a personal quarterly review or a team planning day.</p><p>The toolkit is freely available under Creative Commons (CC BY-NC 4.0). Use it, adapt it, see what grows.</p><div class="kg-card kg-file-card"><a class="kg-file-card-container" href="https://tomcw.xyz/content/files/2026/01/garden-of-ideas--1-.pdf" title="Download" download><div class="kg-file-card-contents"><div class="kg-file-card-title">garden-of-ideas (1)</div><div class="kg-file-card-caption"></div><div class="kg-file-card-metadata"><div class="kg-file-card-filename">garden-of-ideas (1).pdf</div><div class="kg-file-card-filesize">150 KB</div></div></div><div class="kg-file-card-icon"><svg viewbox="0 0 24 24"><defs><style>.a{fill:none;stroke:currentColor;stroke-linecap:round;stroke-linejoin:round;stroke-width:1.5px;}</style></defs><title>download-circle</title><polyline class="a" points="8.25 14.25 12 18 15.75 14.25"/><line class="a" x1="12" y1="6.75" x2="12" y2="18"/><circle class="a" cx="12" cy="12" r="11.25"/></svg></div></a></div><hr><p><em>The Garden of Ideas emerged from years of facilitating strategy sessions with organisations, where good ideas often got lost in the shuffle or the HiPPO (Highest Paid Person&apos;s Opinion) and from my own struggles/hatred with to-do lists that never seemed to shrink. If you use it, I&apos;d genuinely love to hear what worked and if you adapt it, pay it forward with Open Licensing!</em></p>]]></content:encoded></item><item><title><![CDATA[Building Open Recommendations]]></title><description><![CDATA[A full breakdown of how I built Open Recommendations, the inner workings, the choices I made, including provider agnostic AI, strict data structures, deciding whats a recommendation and whats just an observation, RAG, and weighting community feedback.]]></description><link>https://tomcw.xyz/building-open-recommendations/</link><guid isPermaLink="false">68bbd27a2174e13db933b720</guid><category><![CDATA[ai]]></category><category><![CDATA[Data]]></category><category><![CDATA[Open Infrastructure]]></category><dc:creator><![CDATA[Tom Watson]]></dc:creator><pubDate>Sun, 25 Jan 2026 21:49:29 GMT</pubDate><media:content url="https://tomcw.xyz/content/images/2026/01/open-recs-architecture.pptx-2.png" medium="image"/><content:encoded><![CDATA[<img src="https://tomcw.xyz/content/images/2026/01/open-recs-architecture.pptx-2.png" alt="Building Open Recommendations"><p>I&apos;m not sure if you&apos;ve noticed, but there is a lot of buzz around AI; why you should/must use it, how it can save time/do everything, how it will destroy the world. So let&apos;s get this out the way first.</p><p>Yes I will be talking about AI, but there will be no hype or hyperbole here. I will be explaining what and how we&apos;ve used AI in building Open Recommendations as a way of explaining how I think about the use cases, the advantages and disadvantages of certain approaches, and how I approach it from a data point of view. It will be practical and a bit technical but it will be grounded I hope. </p><h2 id="what-is-open-recommendations">What is Open Recommendations?</h2><p>It&apos;s designed to be a central hub to upload, analyse, and track reports and recommendations across the social purpose sector. In its most simple form it:</p><ul><li>allows users to upload a &apos;source&apos; such as a report via either a document or url</li><li>extracts the source into a machine readable format</li><li>summarises and categorises against a taxonomy</li><li>pulls specific recommendations from the source</li><li>categorises these recommendations against a taxonomy</li><li>allows the searching, exploration and chatting with the knowledge base created</li><li>allows community based tracking of action towards recommendations</li></ul><p>Or in more simple terms:</p><ul><li>Get stuff out of documents</li><li>Make use of this knowledge</li><li>Track whether we actually do anything about all the recommendations we make</li></ul><figure class="kg-card kg-image-card"><img src="https://tomcw.xyz/content/images/2026/01/open-recs-architecture.pptx--1-.png" class="kg-image" alt="Building Open Recommendations" loading="lazy" width="960" height="540" srcset="https://tomcw.xyz/content/images/size/w600/2026/01/open-recs-architecture.pptx--1-.png 600w, https://tomcw.xyz/content/images/2026/01/open-recs-architecture.pptx--1-.png 960w" sizes="(min-width: 720px) 720px"></figure><p>You can read more about the why of Open Recommendations <a href="https://www.openrecommendations.com/about?ref=tomcw.xyz" rel="noreferrer">here</a>.</p><h2 id="making-use-of-ai">Making use of AI</h2><p>So yeah I use LLMs in Open Recommendations and what I wanted to do here is break down how I&apos;ve used them, as maybe a helpful guide to others if they are thinking about exploring this stuff. As mentioned there is a lot of stuff written about AI and I find that in the social purpose sector especially we give a very minimal amount of actual detail of what this actually involves. Or if we do give detail it is very technical. I also want to lay out some of the choices I&apos;ve made along the way and why I made them. So I&apos;m going to try to strike a balance here between giving enough detail without getting too technical.</p><p>As I broke down earlier, there are various things that Open Recommendations does. From an AI point of view this can be explored in two main parts: getting stuff in, and making use of the stuff once it&apos;s in.</p><h2 id="ai-ingestion-getting-stuff-in">AI ingestion (getting stuff in)</h2><p>So one of the things we think is useful on Open Recs is the ability to upload a source in the form of a document, which in many cases is a pdf. Now I could write a whole post about the dreaded pdf, but let&apos;s face it, it&apos;s still the most used format for documents. We also allow users to point to a url if they have a source that is HTML which is a much simpler process, so we&apos;ll leave that for now.</p><p>When a user uploads a pdf we do the following:</p><ul><li>Firstly we store that pdf as is</li><li>Then we use an OCR to fully extract everything in it - text, data tables, links - into a machine readable format (markdown) and store this making it possible to fully recreate the pdf but in a more usable format.</li><li>Once we have the full extract we use an LLM to create a summary of the source and categorise it (by source type, purpose, thematic area, role relevance) and store these</li><li>We then create an embedding of this information</li><li>We then extract specific recommendations contained within the source</li><li>And then we categorise those recommendations (purpose, target audience, thematic area, location scope)</li><li>Finally we create an embedding of each recommendation</li></ul><figure class="kg-card kg-image-card"><img src="https://tomcw.xyz/content/images/2026/01/open-recs-architecture.pptx-1.png" class="kg-image" alt="Building Open Recommendations" loading="lazy" width="960" height="540" srcset="https://tomcw.xyz/content/images/size/w600/2026/01/open-recs-architecture.pptx-1.png 600w, https://tomcw.xyz/content/images/2026/01/open-recs-architecture.pptx-1.png 960w" sizes="(min-width: 720px) 720px"></figure><p>So there&apos;s a lot going on here, but to a user it&apos;s simply click upload and await the return to check and make any edits if they choose to. I chose to split these processes up for a couple of reasons.</p><p>Firstly from a technical point of view this allows independent scaling - if recommendation extraction is slow, we can optimise that without touching the OCR step. But more importantly, it means we can swap components in and out. This becomes really important when we talk about AI models.</p><h2 id="the-dont-get-too-attached-to-your-ai-provider-principle">The &quot;don&apos;t get too attached to your AI provider&quot; principle</h2><p>So, in case you hadn&apos;t noticed, AI providers are not stable partners. Pricing changes, models get deprecated, service goes down, and suddenly that OpenAI dependency you built your whole system around becomes a problem.</p><p>So I built Open Recommendations to be provider-agnostic from day one. It currently support over 19 different AI models through a unified interface - Mistral, OpenAI, Anthropic (Claude), Google Gemini, and even some more niche options like GreenPT for those concerned about environmental impact.</p><p>The way this works is fairly straightforward. We have a single function that takes a model name and returns the right provider configured correctly. When we need to make an AI call anywhere in the system, we call this function rather than directly calling OpenAI or whoever. If tomorrow we decide Claude is better for a particular task, we change one configuration value. If Mistral doubles their prices, we switch to Gemini. No rewriting of business logic required.</p><p>This isn&apos;t just about hedging bets though. Different models are genuinely better at different things, and this architecture lets us be intentional about that.</p><figure class="kg-card kg-image-card"><img src="https://tomcw.xyz/content/images/2026/01/open-recs-architecture.pptx--2-.png" class="kg-image" alt="Building Open Recommendations" loading="lazy" width="960" height="540" srcset="https://tomcw.xyz/content/images/size/w600/2026/01/open-recs-architecture.pptx--2-.png 600w, https://tomcw.xyz/content/images/2026/01/open-recs-architecture.pptx--2-.png 960w" sizes="(min-width: 720px) 720px"></figure><h2 id="different-models-for-different-jobs">Different models for different jobs</h2><p>This is where it gets interesting. Not all AI tasks are created equal, and treating them as such is both wasteful and often produces worse results.</p><p>Take our OCR step. When we&apos;re extracting text from complex PDFs - ones with tables, multiple columns, embedded images - we need something that can actually see the document. So we use vision-capable models here, specifically Mistral Vision or Google Vision depending on the document complexity. A pure text model simply can&apos;t do this job.</p><p>But when we&apos;re summarising that extracted text? We don&apos;t need vision capabilities. We need something good at understanding context and producing coherent summaries. Here a model like Claude or GPT-4 Turbo excels.</p><p>For the recommendation extraction, we need something that can follow fairly complex instructions reliably - identifying what is and isn&apos;t actually a recommendation (more on this shortly), categorising against multiple taxonomies, and outputting in a very specific format. This is where model choice really matters, because consistency is everything.</p><p>And for the embedding generation - turning text into numerical vectors for search - we use a completely different type of model altogether. These embedding models are specifically trained for this task and are much cheaper to run than the big language models.</p><p>The point is: matching the model to the task saves money, improves results, and means you&apos;re not paying GPT-4 prices to do work that a smaller model handles perfectly well.</p><h2 id="making-ai-play-nice-with-your-data-strict-structures">Making AI play nice with your data (strict structures)</h2><p>The most consistent thing about LLM&apos;s is that they are not consistent,  they&apos;re unpredictable. Ask the same question twice, you might get differently formatted answers. Ask it to return JSON and it might wrap it in markdown code blocks. Or not. Or add a helpful introduction before the JSON that breaks your parser.</p><p>This is a massive problem when you&apos;re building a system that needs to actually do something with the AI&apos;s output. We need recommendations in a specific format so we can store them in our database, search them, categorise them. &quot;Close enough&quot; doesn&apos;t cut it.</p><p>So we use something called <strong>schema validation</strong> to define exactly what the AI&apos;s output must look like. When the AI returns something, we validate it against this schema. If it doesn&apos;t match, we reject it and try again (or handle the error gracefully).</p><p>For example, a recommendation must have a recommendation_text (string), a target_audience (from a defined list), a thematic_area (also from a defined list), a confidence_score (High, Medium, or Low), and so on. If the AI decides to get creative and return something different, our system catches it.</p><p>This sounds pedantic but it&apos;s absolutely essential. Without it, you end up with data quality issues that compound over time. One recommendation stored with thematic_area as &quot;health&quot; and another as &quot;Health&quot; and another as &quot;Healthcare&quot; - suddenly your filtering and analytics are broken.</p><p>We also use <strong>controlled vocabularies</strong> for our categorisation. Rather than letting the AI free-text categorise things, we give it a specific list: &quot;The thematic areas are: Health, Education, Environment, Housing, Employment...&quot; and so on. The AI must pick from this list. This means our data is consistent and our filters actually work.</p><figure class="kg-card kg-image-card"><img src="https://tomcw.xyz/content/images/2026/01/open-recs-architecture.pptx--3--1.png" class="kg-image" alt="Building Open Recommendations" loading="lazy" width="960" height="540" srcset="https://tomcw.xyz/content/images/size/w600/2026/01/open-recs-architecture.pptx--3--1.png 600w, https://tomcw.xyz/content/images/2026/01/open-recs-architecture.pptx--3--1.png 960w" sizes="(min-width: 720px) 720px"></figure><h2 id="not-everything-is-a-recommendation">Not everything is a recommendation</h2><p>One of the more interesting challenges was teaching the AI what actually constitutes a recommendation. Turns out a lot of text that sounds recommendation-ish... isn&apos;t.</p><p>Consider these three statements from a typical report:</p><ol><li>&quot;The sector needs significant improvement in data sharing practices&quot;</li><li>&quot;Many organisations struggle to secure long-term funding&quot;</li><li>&quot;Local authorities should establish dedicated funding streams for community organisations within the next 18 months&quot;</li></ol><p>Only the third one is actually a recommendation. The first two are observations or statements of the problem. A recommendation needs to be actionable and prescriptive - it should have someone who needs to do something, and ideally a sense of what and when.</p><p>We train our extraction prompts to look for imperative verbs (establish, develop, implement, create, ensure), clear target audiences, and actual calls to action. We also assign confidence scores - if it looks like a recommendation but we&apos;re not sure, we mark it as low confidence and the user can review.</p><p>This filtering is crucial. Without it, you end up with a database full of &quot;recommendations&quot; that are actually just problems restated. Not useful when you&apos;re trying to track action. </p><p>This training is an ongoing process and we don&apos;t always get it right. What we have noticed is that people write in a lot of different ways and are often not explicit in recommendations. There are probably many reasons for this, but it makes it hard, for both humans and especially non-humans to fully pick the intention.</p><figure class="kg-card kg-image-card"><img src="https://tomcw.xyz/content/images/2026/01/open-recs-architecture.pptx--5-.png" class="kg-image" alt="Building Open Recommendations" loading="lazy" width="960" height="540" srcset="https://tomcw.xyz/content/images/size/w600/2026/01/open-recs-architecture.pptx--5-.png 600w, https://tomcw.xyz/content/images/2026/01/open-recs-architecture.pptx--5-.png 960w" sizes="(min-width: 720px) 720px"></figure><h2 id="making-use-of-the-knowledge-search-and-chat">Making use of the knowledge (search and chat)</h2><p>Once we&apos;ve got all this structured data, the next challenge is making it useful. This is where embeddings come in.</p><p>An embedding is essentially turning a piece of text into a list of numbers - a vector - that represents its meaning. Similar texts will have similar vectors. This means we can do semantic search: find recommendations that are conceptually similar even if they use completely different words.</p><p>So if someone searches for &quot;climate action in schools&quot; we can find recommendations about &quot;environmental education programmes&quot; or &quot;reducing carbon footprint in educational settings&quot; - things that are clearly relevant but don&apos;t contain the exact search terms.</p><p>We combine this with traditional keyword search (sometimes you do want exact matches, like a specific organisation name) and the result is a hybrid search that handles both use cases.</p><p>The chat interface builds on this. When a user asks a question, we:</p><ol><li>Turn their question into an embedding</li><li>Find the most relevant sources and recommendations</li><li>Feed those to an AI as context</li><li>Generate a response that&apos;s grounded in the actual content</li></ol><figure class="kg-card kg-image-card"><img src="https://tomcw.xyz/content/images/2026/01/open-recs-architecture.pptx--4-.png" class="kg-image" alt="Building Open Recommendations" loading="lazy" width="960" height="540" srcset="https://tomcw.xyz/content/images/size/w600/2026/01/open-recs-architecture.pptx--4-.png 600w, https://tomcw.xyz/content/images/2026/01/open-recs-architecture.pptx--4-.png 960w" sizes="(min-width: 720px) 720px"></figure><p>This is called <strong>RAG - Retrieval Augmented Generation</strong> - and it&apos;s the difference between an AI that makes things up and one that gives you answers based on actual evidence. The AI can only draw from the sources and recommendations we&apos;ve given it, and we show the user which sources informed the answer. We think this is why Open Recommendations is better/different than more generalised chat interfaces.</p><h2 id="tracking-progress-do-we-actually-do-anything">Tracking progress (do we actually do anything?)</h2><p>This is perhaps the most important and least glamorous part of Open Recommendations. We make a lot of recommendations in this sector. We commission reports, hold consultations, conduct evaluations. And then... what?</p><p>Open Recommendations allows <strong>community-based progress tracking</strong>. Anyone can add an update to a recommendation: &quot;We&apos;ve started work on this&quot;, &quot;This was implemented in our area&quot;, &quot;This was rejected because...&quot;. Each update captures who made it, when, and crucially - a sentiment score.</p><p>The sentiment scoring was an interesting design challenge. We wanted to capture not just &quot;has this been done or not&quot; but the nuance of progress. Is this update positive movement? Negative? Neutral? We use AI to analyse the update text and assign a sentiment score from 0 to 1, which we then aggregate to give an overall sense of momentum on a recommendation.</p><p>This creates a kind of community evidence base. If ten organisations all report that a particular recommendation is impractical, that&apos;s valuable information. If one area has successfully implemented something, their experience can inform others.</p><p>The challenge here is weighting. A progress update from the organisation named in the recommendation probably carries more weight than one from a random observer. We&apos;re still iterating on how best to represent this, but the principle is that tracking should be community-owned, not gatekept.</p><figure class="kg-card kg-image-card"><img src="https://tomcw.xyz/content/images/2026/01/open-recs-architecture.pptx--6-.png" class="kg-image" alt="Building Open Recommendations" loading="lazy" width="960" height="540" srcset="https://tomcw.xyz/content/images/size/w600/2026/01/open-recs-architecture.pptx--6-.png 600w, https://tomcw.xyz/content/images/2026/01/open-recs-architecture.pptx--6-.png 960w" sizes="(min-width: 720px) 720px"></figure><h2 id="what-ive-learned">What I&apos;ve learned</h2><p>Building Open Recommendations has reinforced a few beliefs I had and challenged others.</p><p>The biggest reinforcement: AI is a tool, not magic. It&apos;s very good at certain things (extracting structured information from unstructured text, understanding semantic similarity, generating contextual responses) and poor at others (being consistent, knowing what it doesn&apos;t know, handling edge cases gracefully). The skill is knowing where to deploy it and where to keep humans in the loop.</p><p>The biggest surprise: how much of the work is not the AI part. <strong>Data modelling, validation, error handling, user interface, privacy controls</strong> - the AI is maybe 20% of the actual system. The other 80% is all the boring stuff that makes it actually usable.</p><p>And the ongoing learning: the right architecture matters enormously. The decision to make the AI layer swappable has already paid for itself multiple times as I&apos;ve optimised costs and performance. If I&apos;d hardcoded OpenAI calls throughout the codebase, we&apos;d be in a much harder position now.</p><h2 id="in-closing">In closing</h2><p>If you&apos;re in the social purpose sector and thinking about building something with AI, I hope this has been useful. The technology is genuinely capable of things that would have been impossible a few years ago. But it needs to be approached with clear thinking about what you&apos;re actually trying to achieve, what the AI can realistically do, and how you&apos;ll handle it when things go wrong.</p><p>Open Recommendations isn&apos;t finished - it&apos;s still being developed and improved. But the foundations are solid because we took time to think through the architecture rather than rushing to bolt AI onto something.</p><p>If you&apos;d like to try it, learn more, or get involved, you can find Open Recommendations at <a href="https://www.openrecommendations.com/?ref=tomcw.xyz">https://www.openrecommendations.com/</a>. And if you&apos;ve got questions about anything I&apos;ve covered here, I&apos;m happy to go deeper - just get in touch.</p>]]></content:encoded></item><item><title><![CDATA[Is the best way to actually make AI work in an organisation to focus on constraints?]]></title><description><![CDATA[Pondering on how we actually need to focus on constraints when everything and anything is possible]]></description><link>https://tomcw.xyz/is-the-best-way-to-actually-make-ai-work-in-an-organisation-to-focus-on-constraints/</link><guid isPermaLink="false">696779e02174e13db933c35d</guid><category><![CDATA[ai]]></category><category><![CDATA[Questions]]></category><dc:creator><![CDATA[Tom Watson]]></dc:creator><pubDate>Wed, 14 Jan 2026 11:40:59 GMT</pubDate><media:content url="https://tomcw.xyz/content/images/2026/01/constraints.png" medium="image"/><content:encoded><![CDATA[<img src="https://tomcw.xyz/content/images/2026/01/constraints.png" alt="Is the best way to actually make AI work in an organisation to focus on constraints?"><p>Is the best way to actually make AI work in an organisation to focus on constraints?</p><p>AI offers so many opportunities that we end up doing a bit of everything&#x2026; and realising the gains from none of it. The menu is endless, so you stare overwhelmed and walk away and get chips. </p><p>Users when landed with a new tool that can do so many things often go in wide ranges - questions or tasks that are too broad, too narrow, too difficult, too easy and then don&apos;t see the value. </p><p>Go back a couple of years and prompt engineering was all the rage as it in some ways provided constraints. But then reasoning models came along and we moved away from that. Which is fine for individuals. But not when it comes to AI in organisations and in services</p><p>Constraints are what turn possibility into progress: </p><ul><li>Pick a service (not a shiny tool).</li><li>Name the job to be done and the users who benefit.</li><li>Define what &#x201C;good&#x201D; looks like (time saved, errors reduced, quality improved). </li><li>Set boundaries: what it is for, what it isn&#x2019;t for, and when a human must step in.</li><li> Build the simplest workflow that makes it repeatable, measurable, and safe.</li></ul><p>All the talk at the minute is to experiment, don&apos;t be left behind. Yes, experiment to find the edges, but then commit to a few well-scoped uses and ship them inside real services.</p><p>It&#x2019;s the same with data: having all the data often tells you nothing. Value comes from purpose, definition, and decision-making constraints.</p><p>Constraints can take many forms </p><ul><li>Purpose constraint: This assistant drafts first versions of X for Y audience. It doesn&#x2019;t approve, decide, or sign off. </li><li>Context constraint: Only use these sources/records. If they&#x2019;re missing, ask for them (or say you can&#x2019;t). </li><li>Output constraint: Use this template, include these fields, keep it to 200 words, match this tone. </li><li>Risk constraint: If the topic touches safeguarding/legal/finance, route to a human or require a second check. </li><li>Quality constraint: State assumptions, flag uncertainty, and link back to the source of truth.</li></ul><p>But mainly they come down to For this, not this. Like this, not like this. </p><p>Hmm. Maybe I&#x2019;ve just described service design.</p>]]></content:encoded></item><item><title><![CDATA[The end of Data For Action]]></title><description><![CDATA[Reflections on building a different kind of consultancy- one that chose uncertainty over certainty, conversations over transactions, and approach over outputs.]]></description><link>https://tomcw.xyz/the-end-of-data-for-action/</link><guid isPermaLink="false">695a7c3c2174e13db933c105</guid><category><![CDATA[Open Working]]></category><category><![CDATA[Random]]></category><dc:creator><![CDATA[Tom Watson]]></dc:creator><pubDate>Sun, 04 Jan 2026 17:18:29 GMT</pubDate><media:content url="https://tomcw.xyz/content/images/2026/01/Copy-of-Brand-Principle-Posters-17--2-.png" medium="image"/><content:encoded><![CDATA[<img src="https://tomcw.xyz/content/images/2026/01/Copy-of-Brand-Principle-Posters-17--2-.png" alt="The end of Data For Action"><p>After three years, Tom (F) and I are bringing Data For Action to a close. As Tom heads off to exciting new pastures, we&apos;re ending it together - Tom &amp; Tom are no more. But before looking to the future, I wanted to capture what we learned, what we built, and what I hope others might take forward.</p><h2 id="what-we-learned-about-working-differently"><strong>What We Learned About Working Differently</strong></h2><p></p><h3 id="we-chose-uncertainty-and-it-made-our-work-better-but-made-it-harder-to-actually-make-money"><strong>We Chose Uncertainty, and It Made Our Work Better (but made it harder to actually make money)</strong></h3><p>One of our core ideals was to <strong>lean into uncertainty</strong>. We didn&apos;t go into projects with preconceived solutions. We asked better questions. We prototyped. We experimented. And yes, this probably cost us some contracts - funders and clients often want the comfort of certainty, especially when budgets are tight. But the work we did produce was deeper, more contextual, and more likely to actually work because it emerged from real conversations with real people.</p><p>One thing I learned through all of our work was this: <em>the best results come from being open to possibilities</em>. Going in with a preconceived idea may be easier to deliver, but it likely won&apos;t be the best outcome. The tension you feel when you don&apos;t know exactly where a workshop will lead? I think that&apos;s a feeling that tells you you&apos;re pushing at the boundaries of what&apos;s possible and I think this is a good place to be.&#xA0;</p><p><strong>What I hope others do differently</strong>: Reward uncertainty. Fund discovery. Value questions as much as answers. The sector&apos;s tendency to demand certainty before releasing resources means we often fund safe, predictable work that doesn&apos;t actually shift the needle.</p><h3 id="data-and-maps-as-conversations-not-just-outputs"><strong>Data and Maps as Conversations, Not Just Outputs</strong></h3><p>We developed a way of thinking that fundamentally changed how we approached our work:<a href="https://tomcw.xyz/data-as-conversations/"> <strong><u>data as conversations</u></strong></a> and<a href="https://tomcw.xyz/maps-as-conversations/"> <strong><u>maps as conversations</u></strong></a>.</p><p>When we worked with people in Sheffield to map their neighbourhoods, this wasn&#x2019;t a technical approach, we weren&apos;t just drawing boundaries - we were creating spaces for people to talk about place, belonging, and what matters to them. We built<a href="https://www.mapmypatch.co.uk/?ref=tomcw.xyz"><u> <strong>Map My Patch</strong></u></a> to support this work, allowing communities to define their own places rather than accepting administrative boundaries that meant nothing to them.</p><p>When we developed the question-based approach for the Insight Infrastructure work, we created<a href="https://www.questionsforaction.com/?ref=tomcw.xyz"> <strong><u>Questions For Action</u></strong></a> - a tool that takes users through our questions-based approach, helping teams move from &apos;what data do we need?&apos; to &apos;what questions are we trying to answer?&apos;</p><p>This shift from static outputs to dynamic conversations meant that our work could evolve, be owned by communities, and continue long after we moved on.</p><p><strong>What I hope others do differently</strong>: Stop treating data and maps as finished products. They&apos;re conversation starters. They&apos;re ways of bringing people together to discuss what matters, what&apos;s changing, and what we might do about it.&#xA0;</p><h2 id="the-work-were-proud-of"><strong>The Work We&apos;re Proud Of</strong></h2><h3 id="sheffield-neighbourhoods-when-communities-own-the-map"><strong>Sheffield Neighbourhoods: When Communities Own the Map</strong></h3><p>This work, led brilliantly by Tom F, demonstrated everything we believed about community-owned data and citizen-led approaches. Sheffield now has 147 citizen-defined neighbourhood boundaries that are actually used in conversations across sectors.</p><p>The innovation here wasn&apos;t technical - it was about power and ownership. We showed that <strong>when people recognise themselves in the data, they engage differently</strong>. They talk about challenges, offer ideas, take ownership, and build resilience.</p><p><strong>What we learned</strong>: People don&apos;t live according to administrative boundaries. They recognise physical markers, relationships, and neighborhoods. Data that reflects this reality is more useful.</p><p><strong>What didn&apos;t work</strong>: While people lauded the approach, almost no one wanted to properly fund it. Huge sums went to commissions writing white papers and holding events at the House of Lords (we declined with a &apos;what a f*cking bizarre place to hold a neighbourhood event&apos; response), while we ran this work on a shoestring.</p><p><strong>What I hope others do differently</strong>: Actually fund the hard, transformative work, not just the comfortable stuff. Citizen-led approaches to power and data take time and sustained investment. They&apos;re messy and uncertain - and that&apos;s exactly why they&apos;re valuable.</p><p><strong>Tools we built</strong>:<a href="https://www.mapmypatch.co.uk/?ref=tomcw.xyz"><u> <strong>Map My Patch</strong></u></a> to support communities in mapping their own places and having conversations about what they see.</p><h3 id="local-needs-databank-flexible-standards-for-the-ai-era"><strong>Local Needs Databank: Flexible Standards for the AI Era</strong></h3><p>With the databank we tried to solve a fundamental problem: how do you create data standards that are strict enough to be useful but flexible enough that people will actually use them?</p><p>Our answer was to use metadata and Schema.org standards to create a middle ground. You could drop in a CSV file, tell us which columns meant what, and we&apos;d handle the rest. Your column could say &apos;location&apos; or &apos;where&apos; or &apos;office&apos; - we didn&apos;t care, we&apos;d make it work.</p><p><strong>What we learned</strong>: Flexible standards and metadata aren&apos;t just nice-to-have - they&apos;re becoming <em>more</em> important as we work in the AI era. The ability to describe and transform data flexibly is crucial.</p><p><strong>What didn&apos;t work</strong>: We focused heavily on the contribution/upload mechanism (rightly, I still believe), but when leadership changed at the client, new priorities emerged. The project never got the continued development it needed. This taught us that organisational memory is long on emotions but short on details - and that sometimes you need to deliver something people can &quot;hang their hat on&quot; early to maintain support through leadership changes.</p><h3 id="insight-infrastructure-questions-before-data"><strong>Insight Infrastructure: Questions Before Data</strong></h3><p>Working with Joseph Rowntree Foundation and charities across the UK, we explored what a <strong>data sharing movement</strong> might look like. But instead of starting with &quot;what data should we share?&quot;, we started with &quot;what questions do we need to answer?&quot;</p><p>This led to developing our <strong>question-based approach</strong> and <strong>question banking</strong> methodology - now available through<a href="https://www.questionsforaction.com/?ref=tomcw.xyz"> <strong><u>Questions For Action</u></strong></a>.</p><figure class="kg-card kg-image-card"><img src="https://tomcw.xyz/content/images/2026/01/Data-for-action-question-approach--2-.jpg" class="kg-image" alt="The end of Data For Action" loading="lazy" width="960" height="604" srcset="https://tomcw.xyz/content/images/size/w600/2026/01/Data-for-action-question-approach--2-.jpg 600w, https://tomcw.xyz/content/images/2026/01/Data-for-action-question-approach--2-.jpg 960w" sizes="(min-width: 720px) 720px"></figure><p><strong>What we learned</strong>:</p><ul><li>Questions are data layers in themselves. They reveal assumptions, clarify priorities, and create dynamic data ecosystems</li><li>Minimum viable data standards can work if governance is strong</li><li>Starting with questions builds ownership and purpose</li></ul><p><strong>What didn&apos;t work</strong>: Again, changes in leadership meant we never got to develop the prototypes further.</p><p><strong>What I hope others do differently</strong>: Start with questions. Always. And capture those questions as data - especially in an AI era, the journey of inquiry is as valuable as the destination.</p><h3 id="clientearth-cultural-change-through-showing-things"><strong>ClientEarth: Cultural Change Through Showing Things</strong></h3><p>Our 18-month engagement with ClientEarth taught us perhaps our most valuable lessons. We summarised them in<a href="https://tomcw.xyz/lessons-learned-from-working-together/"> <strong><u>Lessons Learned from Working Together</u></strong></a>, but a few stand out:</p><ul><li><strong>Show things, don&apos;t just talk about them</strong>. Even imperfect prototypes beat perfect descriptions.</li><li><strong>Vibes matter</strong>. Creating enough trust and safety to be playful and experiment liberates people from stifling professional expectations.</li><li><strong>Clear language beats jargon</strong>. Simple communication ensures everyone knows what you mean - especially important in multilingual environments.</li><li><strong>The internal team makes it work</strong>. No matter how good we were, ClientEarth&apos;s team made this a success by trusting us and the process.</li></ul><p><strong>What we learned</strong>: Approach matters more than outputs. If you don&apos;t bring people along with you, the final tool won&apos;t matter. When the team just started <em>using</em> the impact tool without fanfare, we knew we&apos;d got it right.</p><h3 id="wildlife-trusts-getting-prototyping-right"><strong>Wildlife Trusts: Getting Prototyping Right</strong></h3><p>With the Wildlife Trusts, we hit our stride. We translated their Evidence Competency Framework into a self-assessment tool, and crucially, everyone understood it was a <em>prototype</em>. There was no scope creep. We did enough to validate assumptions and support the team to secure funding for full development.</p><p><strong>What we learned</strong>: When expectations are clear and everyone embraces where we are and what we are really trying to achieve, work stays focused and valuable. Sometimes prototypes are there to show you what not to take forward as much as what to take forward. </p><h2 id="what-we-built-that-will-keep-working"><strong>What We Built That Will Keep Working</strong></h2><p>Beyond specific client projects, we developed tools and approaches that are now available for the sector:</p><ul><li><a href="https://www.questionsforaction.com/?ref=tomcw.xyz"><strong><u>Questions For Action</u></strong></a> - Our question-based prioritisation tool for individuals and teams</li><li><a href="https://www.openrecommendations.com/?ref=tomcw.xyz"><strong><u>Open Recommendations</u></strong></a> - Nearly 200 sources of sector recommendations, AI-parsed and searchable</li><li><a href="https://www.mapmypatch.com/?ref=tomcw.xyz"><strong><u>Map My Patch</u></strong></a> - Community-led neighbourhood mapping tools</li><li><strong>Sheffield Neighbourhoods Archive</strong> -<a href="https://dataforaction-tom.github.io/sheffield-neighbourhoods-boundaries/?ref=tomcw.xyz"> <u>Boundary files</u></a> preserved for the long term</li><li><strong>Question banking methodology</strong> - <a href="https://dataforaction.notion.site/Prototyping-insight-infrastructure-for-the-charity-sector-b53e4b066c2440f6b91f1ad0f334fc8c?pvs=74&amp;ref=tomcw.xyz"><u>Available through our Notion repository</u></a></li><li><strong>Flexible data standards approach</strong> - Documented in our <a href="https://www.notion.so/dataforaction/Local-Needs-Databank-9da5de58ca65430187e6c62f610ba0a8?ref=tomcw.xyz"><u>project repositories</u></a></li></ul><h2 id="reflections-on-running-a-small-different-agency"><strong>Reflections on Running a Small, Different Agency</strong></h2><p>Were we really an agency? Probably not. When you worked with us, you got Tom and Tom. We were purposely small because we wanted to do good, important, deep work, not scale a business.</p><p>This left us in an unusual position. We had no junior staff to pad margins, but we weren&apos;t just freelancers either. We took risks. We tried to change how people did things. We embraced uncertainty when others wanted certainty.</p><p><strong>What I learned</strong>: Being different has costs. As the cost-of-living crisis hit, people took fewer risks. Work we pitched for - like Local Motion&apos;s learning framework, Place Matters mapping, Ealing&apos;s community asset mapping - we might have secured if we&apos;d been more traditional, more certain, more safe in our proposals.</p><p>But that&apos;s not how I work. And I don&apos;t regret it.</p><p><strong>What I hope others do differently</strong>: Value the different approaches. Fund the teams doing things differently. Pay them properly for the risk and innovation they bring. Don&apos;t just say you value innovation - actually resource it.</p><h2 id="on-being-open-by-default"><strong>On Being Open By Default</strong></h2><p>We actively shared our work, opened up our processes, and showed the inner workings. I think this will have an ongoing impact. But I also think people assumed we were somehow funded to do this - that money was readily available for this kind of sharing. It wasn&apos;t.</p><p><strong>What I learned</strong>: Open working is valuable and rare - but it needs to be recognised and funded accordingly. Transparency and open documentation should be valued, not taken for granted.</p><h2 id="building-resilience-through-uncertainty"><strong>Building Resilience Through Uncertainty</strong></h2><p>One of my core frameworks - developed more fully in my solo work - is about <strong>collective resilience</strong>: the capacity of interconnected organisations to anticipate, prepare for, respond to, and adapt collectively to change, disruption, and uncertainty.</p><p>Data For Action embodied this. We built tools and approaches that helped organisations embrace uncertainty rather than avoid it. We created structures that liberated rather than constrained. We asked questions that opened up possibilities rather than closed them down.</p><p><strong>What I hope others do differently</strong>:</p><ul><li>Build for resilience, not just efficiency</li><li>Create space for uncertainty and experimentation</li><li>Trust that the messy, conversational, question-led approach might be harder to fund but is more likely to create lasting change</li><li>Remember that approach matters as much as outputs</li></ul><h2 id="what-im-taking-forward"><strong>What I&apos;m Taking Forward</strong></h2><p>I&apos;m immensely proud of what we&apos;ve done over the last three years. We might not have got everything right, but we did it the right way. We embraced uncertainty, we were open by default, we really <em>lived</em> our principles.</p><p>The tools we built - Questions For Action, Open Recommendations, Map My Patch - continue to be available. The approaches we developed - data as conversations, maps as conversations, question-based methodology - are documented and ready for others to adapt. The citizen led boundary files for 147 neighbourhoods in sheffield are available on an archive site&#xA0;</p><p>And this principle? Well obviously this is one I keep:</p><figure class="kg-card kg-image-card"><img src="https://tomcw.xyz/content/images/2026/01/Copy-of-Brand-Principle-Posters-23-1.png" class="kg-image" alt="The end of Data For Action" loading="lazy" width="2000" height="2828" srcset="https://tomcw.xyz/content/images/size/w600/2026/01/Copy-of-Brand-Principle-Posters-23-1.png 600w, https://tomcw.xyz/content/images/size/w1000/2026/01/Copy-of-Brand-Principle-Posters-23-1.png 1000w, https://tomcw.xyz/content/images/size/w1600/2026/01/Copy-of-Brand-Principle-Posters-23-1.png 1600w, https://tomcw.xyz/content/images/size/w2400/2026/01/Copy-of-Brand-Principle-Posters-23-1.png 2400w" sizes="(min-width: 720px) 720px"></figure><p>I&apos;ll miss working with Tom F, but I&apos;m excited to see what he does next. And I&apos;m excited about what comes next for me, carrying forward everything we learned about working differently, embracing uncertainty, and creating conversations that matter.</p><p><em>Want to learn more about our approach or use our tools? Everything is documented and available:</em></p><ul><li><a href="https://www.questionsforaction.com/?ref=tomcw.xyz"><em><u>Questions For Action</u></em></a><em> - Prioritise your most important questions</em></li><li><a href="https://www.openrecommendations.com/?ref=tomcw.xyz"><em><u>Open Recommendations</u></em></a><em> - Search sector recommendations</em></li><li><a href="https://www.mapmypatch.co.uk/?ref=tomcw.xyz"><em><u>Map My Patch</u></em></a><em> - Community-led mapping tool</em></li><li><a href="https://tomcw.xyz/data-as-conversations/"><em><u>Data as Conversations</u></em></a><em> - Our thinking on community-owned data</em></li><li><a href="https://tomcw.xyz/maps-as-conversations/"><em><u>Maps as Conversations</u></em></a><em> - Our approach to citizen-led mapping</em></li><li><a href="https://dataforaction.org.uk/?ref=tomcw.xyz"><em><u>Project repositories</u></em></a><em> - All our documented work</em></li></ul>]]></content:encoded></item><item><title><![CDATA[My Memory - A startup idea for 2026]]></title><description><![CDATA[One of a number of ideas I'm thinking about ahead of 2026. The personal context passport lets people carry a portable, user-owned profile they can share on their terms - from low-stakes convenience (travel, fitness) and evolving into granular, time-limited permissions for high-stakes public services]]></description><link>https://tomcw.xyz/my-memory-a-startup-idea-for-2026/</link><guid isPermaLink="false">693c953e2174e13db933bec2</guid><category><![CDATA[Open Infrastructure]]></category><category><![CDATA[Data]]></category><category><![CDATA[Systems]]></category><category><![CDATA[Tech]]></category><dc:creator><![CDATA[Tom Watson]]></dc:creator><pubDate>Wed, 17 Dec 2025 17:23:40 GMT</pubDate><media:content url="https://images.unsplash.com/photo-1533158307587-828f0a76ef46?crop=entropy&amp;cs=tinysrgb&amp;fit=max&amp;fm=jpg&amp;ixid=M3wxMTc3M3wwfDF8c2VhcmNofDJ8fG1lbW9yeXxlbnwwfHx8fDE3NjU5OTE2OTR8MA&amp;ixlib=rb-4.1.0&amp;q=80&amp;w=2000" medium="image"/><content:encoded><![CDATA[<h3 id="solving-the-%E2%80%9Ccold-start%E2%80%9D-with-a-personal-context-passport">Solving the &#x201C;cold start&#x201D; with a personal Context Passport</h3><img src="https://images.unsplash.com/photo-1533158307587-828f0a76ef46?crop=entropy&amp;cs=tinysrgb&amp;fit=max&amp;fm=jpg&amp;ixid=M3wxMTc3M3wwfDF8c2VhcmNofDJ8fG1lbW9yeXxlbnwwfHx8fDE3NjU5OTE2OTR8MA&amp;ixlib=rb-4.1.0&amp;q=80&amp;w=2000" alt="My Memory - A startup idea for 2026"><p>If I was going to build a startup in 2026 (and I might) here&apos;s one idea I would be looking at:&#xA0; solving <strong>The Cold Start Problem </strong>with a personal context passport<strong>.</strong></p><p>In commercial AI, the &#x201C;<a href="https://en.wikipedia.org/wiki/Cold_start_(recommender_systems)?ref=tomcw.xyz"><u>cold start</u></a>&#x201D; is that moment of friction where the machine knows nothing about you. When you load up a new app or platform you have to paste in your preferences, your history, and your constraints. Again. And again. Think about the amount of times you have to fill information in when searching for a holiday.&#xA0;</p><p>But what if we flipped the model?</p><p>What if, instead of every app building (and owning) a profile of you, you held a portable context that you could share on your terms?</p><p>This idea starts as a way of making everyday things, like shopping, travel, and fitness easier. But it could end somewhere much bigger: giving people control over their data in the most critical moments of their lives.</p><h3 id="version-1-low-fi-the-me-file"><strong>Version 1 (low-fi): The &quot;Me&quot; File </strong></h3><p>Right now, we could sort of solve the cold start problem with simple documents.&#xA0;For the sake of the investors we&apos;ll call this a &apos;Context Passport&apos;, but really it&apos;s just a simple, structured document you use when you interact with LLMs. Set some context fields, set your preferences, and away you go. So next time you interact with a travel &#x2018;agent&#x2019; (as in an AI Agent - see what I did there) or a fitness coach LLM, you don&apos;t need to type out your history. You just upload the relevant &#x2018;passport&#x2019;.</p><h4 id="here%E2%80%99s-an-example-the-travel-context"><strong>Here&#x2019;s an example: The Travel Context</strong></h4><p>Planning a trip is a pain. Yes there are apps that pull together bits of this, but you have to constantly jumped around, re-entering preferences or being stuck with one vendor. And it&#x2019;s why agents are being pushed in this space to make life easier&#x2026; but soon there will be a proliferation of different agents and then shortly after you&#x2019;ll be answering the same twenty questions about budgets and preferences. Could a travel passport solve this? Maybe it looks something simple like this?</p><h1 id="travel-contextyour-name">Travel Context - [Your Name]</h1>
<h2 id="logistics">Logistics</h2>
<ul>
<li><strong>Home Airport</strong>: NCL (Newcastle)</li>
<li><strong>Local Train Station</strong>: Newcastle</li>
<li><strong>Citizenship</strong>: UK</li>
<li><strong>Loyalty</strong>: British Airways (Gold)</li>
</ul>
<h2 id="train-preferences">Train Preferences</h2>
<ul>
<li><strong>Seat</strong>: Table, quiet carriage, window</li>
</ul>
<h2 id="accommodation-style">Accommodation Style</h2>
<ul>
<li><strong>Vibe</strong>: Boutique/Independent over large chains.</li>
<li><strong>Must-haves</strong>: High-speed WiFi (remote work), blackout curtains.</li>
<li><strong>Avoid</strong>: All-inclusive resorts, ground floor rooms.</li>
</ul>
<h2 id="travel-pace">Travel Pace</h2>
<ul>
<li><strong>Style</strong>: &quot;Deep Dive&quot; - prefer 5 days in one city vs. 5 cities in 5 days.</li>
<li><strong>Food</strong>: Priority on street food/local markets over fine dining.</li>
</ul>
<h4 id="here%E2%80%99s-another-example-the-fitness-context"><strong>Here&#x2019;s another example: The Fitness Context</strong></h4><p>The fitness industry is huge and the push for AI coaches is big business. But most are too general to really help you as they don&#x2019;t have the <em>context of you</em>. So what about if you held that context and shared it with your new &#x2018;coach&#x2019; allowing the AI to offer advice that might be closer to your needs, preferences and constraints.&#xA0;</p><h1 id="fitness-wellbeing-contextyour-name">Fitness &amp; Wellbeing Context - [Your Name]</h1>
<h2 id="physiological-profile">Physiological Profile</h2>
<ul>
<li><strong>Injuries</strong>: Left knee ACL reconstruction (2019) - avoid high impact.</li>
<li><strong>Conditions</strong>: Mild asthma (triggered by cold air).</li>
</ul>
<h2 id="goals">Goals</h2>
<ul>
<li><strong>Primary</strong>: Functional strength for hiking.</li>
<li><strong>Anti-Goal</strong>: Not interested in bodybuilding or aesthetics/weight loss focus.</li>
</ul>
<h2 id="logistics">Logistics</h2>
<ul>
<li><strong>Equipment Access</strong>: Kettlebells (12kg, 16kg), Yoga mat. No gym membership.</li>
<li><strong>Time Constraints</strong>: 30 mins max, Mon-Fri mornings.</li>
</ul>
<p>Now of course at this stage, this is just a file you control, and in fact this is something you could make yourself and sort of do this process now, by adding it into any of the chat interfaces, store it in chatGPT as context, add it as artefact in Claude. But you need to do this for every interface, and if we really want to tackle the cold start problem we need to control our data. And as these files grow, simply &quot;pasting&quot; them isn&apos;t safe or efficient. You don&apos;t want the retail bot to see your health data, or the travel bot to see your housing history.</p><h3 id="version-2-the-logic-layer-granular-permissions"><strong>Version 2: The Logic Layer (Granular Permissions)</strong></h3><p>This is where the opportunity shifts from &#x201C;file management&#x201D; to infrastructure.</p><p>If we want portable context to work across tools, we need a way to grant specific apps access to <strong>specific slices</strong> of our context, for a <strong>specific purpose</strong>, for a <strong>limited time</strong>, with <strong>revocation</strong> and <strong>audit logs</strong> built in.</p><p>Protocols like SOLID point to an architecture where your data lives in a personal pod (user-controlled), not a corporate silo.</p><p>Here&#x2019;s what that kind of permission layer might look like:</p><p>section: housing_context.current_situation<br>
access_control:<br>
&#xA0;&#xA0;- service: &quot;newcastle_housing&quot;<br>
&#xA0;&#xA0;&#xA0;&#xA0;granted: &quot;2025-11-02T10:30:00Z&quot;<br>
&#xA0;&#xA0;&#xA0;&#xA0;expires: &quot;2026-06-01T23:59:59Z&quot;<br>
&#xA0;&#xA0;&#xA0;&#xA0;scope: &quot;full_read&quot;<br>
&#xA0;&#xA0;- service: &quot;housing_provider&quot;<br>
&#xA0;&#xA0;&#xA0;&#xA0;granted: &quot;2025-11-10T14:15:00Z&quot;<br>
&#xA0;&#xA0;&#xA0;&#xA0;expires: &quot;2026-03-15T23:59:59Z&quot;<br>
&#xA0;&#xA0;&#xA0;&#xA0;scope: &quot;full_read&quot;<br>
audit:<br>
&#xA0;&#xA0;last_accessed: &quot;2026-01-14T14:22:18Z&quot;<br>
&#xA0;&#xA0;access_count: 23<br>
&#xA0;&#xA0;accessed_by: [&quot;newcastle_housing&quot;]</p>
<p>When a service requests your context:</p><ol><li><strong>Check:</strong> The system verifies the service&apos;s ID against your permission block.</li><li><strong>Filter:</strong> It strips out everything except the scoped data (e.g., Housing sees tenancy history, but <em>not</em> your fitness goals).</li><li><strong>Log:</strong> It records exactly who looked at your data and when.</li></ol><h3 id="version-3-the-integration-layer"><strong>Version 3: The Integration Layer</strong></h3><p>For consumer products, the cold start is annoying. For essential public services, it is worse and possibly actively harmful.</p><p>Currently, a person who interacts with public services often repeats their story across different agencies. <em>They</em> act as the &quot;integration layer,&quot; carrying their context from the housing office to the food bank to the social worker.</p><p>By applying the same logic we used for the &quot;Travel Passport&quot; - portable, user-owned context maybe we can revolutionise public services for the better.</p><p>Imagine that you controlled a portable memory vault. Your context, your history, your needs, your circumstances, your preferences, owned by you. When you need to access a service, you grant it temporary, revocable access to the specific parts of your context it needs to help you. Not some monolithic centralised government database that knows everything about everyone (or tries to). Not fragmented silos that know nothing about each other.</p><p>Citizen-owned, portable context that you carry with you and share on your terms.</p><p>Imagine Alex is trying to stabilise a housing situation: rent arrears have built up after a job change, and three separate processes are in motion:</p><ul><li>The <strong>council housing team</strong> is assessing homelessness prevention support.</li><li><strong>Universal Credit</strong> needs updated rent and landlord details.</li><li>An <strong>energy supplier</strong> is discussing a payment plan and whether Alex should be treated as needing extra support on the account.</li></ul><p>Alex shouldn&#x2019;t have to re-upload the same documents and re-explain the same timeline three times.</p><p>Instead, Alex opens their Context Passport and creates a <strong>time-limited &#x201C;Housing Pack&#x201D;</strong>.</p><p>It contains only:</p><ul><li>Tenancy basics (address, start date, landlord contact)</li><li>Rent amount + arrears summary (with a supporting statement/letter)</li><li>Household basics (optional and minimal)</li><li>Preferred contact method + safe times to call</li></ul><p>A short, user-written note: <em>what&#x2019;s changed, and what Alex is asking for</em><br>Then Alex shares different slices with different services:</p><ul><li><strong>Council housing team:</strong> tenancy + arrears summary + the note (valid for 30 days)</li><li><strong>Universal Credit:</strong> rent amount + landlord details only (valid for 14 days)</li><li><strong>Energy supplier:</strong> support/payment-plan eligibility flag + contact preferences (valid for 7 days)</li></ul><p>No service gets a full profile. Nothing is &#x201C;always on&#x201D;. Shares expire by default.</p><h3 id="how-it-works-in-practice"><strong>How it works in practice</strong></h3><ul><li><strong>Request:</strong> the council housing team requests: &#x201C;We need tenancy details, arrears amount, and confirmation of current income support.&#x201D;</li><li><strong>Consent:</strong> Alex&#x2019;s wallet shows the request in plain language and offers a minimum-share option (e.g., share a UC award statement + arrears letter rather than full bank statements).</li><li><strong>Coordination:</strong> with permission, Universal Credit and/or the landlord provides a simple verification back into Alex&#x2019;s context: &#x201C;UC housing element in payment&#x201D; / &#x201C;arrears total confirmed as &#xA3;X on date Y.&#x201D; The housing team can view <em>that</em> without Alex acting as courier.</li><li><strong>Revocation:</strong> shares expire automatically; Alex can revoke access instantly if circumstances change.</li></ul><p>Less repetition, fewer missed details, fewer &#x201C;prove it again&#x201D; loops.</p><p><strong>The Startup Opportunity</strong></p><p>So we even have some companies starting to look at this for specific use cases, companies like <a href="https://mem0.ai/?ref=tomcw.xyz"><u>Mem0</u></a> , and there is talk of developing an Open Context Layer (OCL). We have existing technology (SOLID, encrypted pods, LLMs). We have layers for verifiable credentials for identity and things like Open Badges. But at the moment it feels like there is a gap in the personal context layer in the commercial space and especially in the citizen space.&#xA0;</p><p>For me the &#x2018;startup&#x2019; opportunity is building the citizen-facing layer that makes this actually usable for public services.</p><p><strong>The Citizen Data Wallet</strong>: A personal vault where you control your own data. Health records, housing history, educational credentials, benefit assessments, support needs, outcomes from previous interventions. Not scattered across 40 different systems, but in one place you control.</p><p><strong>Granular Permission Management</strong>: When you access a service, you grant it read access to specific parts of your context. The housing team needs your current situation and support history. They don&apos;t need your health records. You grant specific access, time-limited, revocable at any point.</p><p><strong>The Service Integration Layer</strong>: Public services connect to a standard protocol, not to your specific wallet directly. They request: &quot;we need context about housing need and family composition to assess your application.&quot; Your wallet asks: &quot;are you okay sharing this?&quot; You approve. The service gets what it needs, nothing more.</p><p><strong>Cross-Service Context Building</strong>: When you receive support from multiple services, they can (with your permission) contribute to your shared context. The mental health service logs that you&apos;ve completed a trauma-informed programme. The housing service can see this (if you grant access), understanding you have support in place.&#xA0;</p><h3 id="why-this-might-be-useful"><strong>Why this might be useful</strong></h3><p>We are moving toward a world of AI agents. If we don&apos;t build this infrastructure, there is a risk that every government department and company will build their own &quot;profile&quot; of you, or there is a massive centralised data platform - eek.&#xA0; You won&apos;t know what&apos;s in it, you won&apos;t be able to correct it, and you won&apos;t be able to take it with you.</p><p>By starting with simple, personal context, like shopping, travel, and fitness maybe we normalize the idea that <strong>you own your data</strong>. Once that pattern is established, perhaps we can scale it to solve the &quot;explaining who I am and my context 15 times to different services&quot; problem, turning a bureaucratic nightmare into a system of dignity and control.</p>]]></content:encoded></item><item><title><![CDATA[Predictions for 2026 (or: What I'm Actually Thinking About)]]></title><description><![CDATA[The Predictions for 2026 that no-one asked for.  Some of them are real, some are more just hope, most of them are firmly tongue in cheek. 

Also I set out 8 ideas I'm thinking about when it comes to tech and AI infrastructure. ]]></description><link>https://tomcw.xyz/predictions-for-2026-or-what-im-actually-thinking-about/</link><guid isPermaLink="false">693c9b9e2174e13db933bedb</guid><category><![CDATA[ai]]></category><category><![CDATA[Data]]></category><category><![CDATA[Open Infrastructure]]></category><category><![CDATA[Tech]]></category><dc:creator><![CDATA[Tom Watson]]></dc:creator><pubDate>Sat, 13 Dec 2025 14:35:23 GMT</pubDate><media:content url="https://images.unsplash.com/photo-1546188994-07c34f6e5e1b?crop=entropy&amp;cs=tinysrgb&amp;fit=max&amp;fm=jpg&amp;ixid=M3wxMTc3M3wwfDF8c2VhcmNofDF8fHRoZSUyMGZ1dHVyZXxlbnwwfHx8fDE3NjU2MzU1MTh8MA&amp;ixlib=rb-4.1.0&amp;q=80&amp;w=2000" medium="image"/><content:encoded><![CDATA[<img src="https://images.unsplash.com/photo-1546188994-07c34f6e5e1b?crop=entropy&amp;cs=tinysrgb&amp;fit=max&amp;fm=jpg&amp;ixid=M3wxMTc3M3wwfDF8c2VhcmNofDF8fHRoZSUyMGZ1dHVyZXxlbnwwfHx8fDE3NjU2MzU1MTh8MA&amp;ixlib=rb-4.1.0&amp;q=80&amp;w=2000" alt="Predictions for 2026 (or: What I&apos;m Actually Thinking About)"><p>It&apos;s December 2025, and honestly I&apos;m tired. Not in a bad way. Though my running has gone to shit recently, I quit drinking and found no noticeable benefit, it&apos;s dark and I just spent two days in a microsoft environment trying to just do, well, anything. Ok, maybe in a bad way...</p><p>I&apos;m thinking about a lot of things recently. The cheese toastie van, obviously. But I&apos;m also considering where I want to put my energy next year and yes I&apos;m thinking about the world and technology and AI. Because of course I am. Everyone is.  </p><p>But, before we get into the bigger ideas here are my totally unasked for predictions for 2026. Some of them are real, some are more just hope, most of them are firmly tongue in cheek. </p><h2 id="the-predictions-nobody-needs">The Predictions Nobody Needs</h2><p><strong>Prediction 1</strong>: Someone, probably a senior leader or a trustee, will ask you about quantum computing. I fucking guarantee it. And you should ignore them. Unless obviously you are working on something so specific that only quantum computing will help you, like climate change modelling. </p><p><strong>Prediction 2</strong>: Community will matter more than audience. The era of trying to reach a million people is dying. In 2026, you&apos;ll realise that having a group chat of 15 people who actually give a shit about you is worth more than 10k LinkedIn followers who are just bots autofarming engagement.</p><p><strong>Prediction 3</strong>: Pizza will still be my favourite food and I will perfect it (I won&apos;t).</p><p><strong>Prediction 4</strong>: The best technical solution you implement will be something incredibly boring like &quot;we standardised our file naming convention&quot; and it will save more time than any AI tool.</p><p><strong>Prediction 5</strong>: Authentic voice will begin to rise up again as everyone just gets fucking sick of AI slop. People want to feel something. <a href="https://youtu.be/nqww88x584U?si=jp5vu89Tq9Iw5Nv8&amp;ref=tomcw.xyz" rel="noreferrer">They want to throw their computer out the window and make a Yaz record.</a> They want to read things written by humans who are tired, who aren&apos;t sure, who don&apos;t sanitise. Not perfectly optimised, SEO-friendly, AI-generated content that says nothing in 1000 words.</p><p><strong>Prediction 6</strong>: There will be some major AI discrimination/shitstorm in the UK related to public services. The organisation won&apos;t have kept proper logs of how decisions were made, because the model is not open so they can&apos;t and there wasn&apos;t anything in place to really evaluate decisions of agents. This will trigger a wave of &quot;oh shit, we need evaluation infrastructure&quot; panic.</p><p><strong>Prediction 7</strong>: Microsoft will release a new set of guidance and documentation that will not make any sense. In fact, hard as it is to believe, it will make less sense than before.</p><p><strong>Prediction 8</strong>: People will lean more and more into early 2000s tech. iPods. <a href="https://aftermath.site/digita-audio-player-snowsky-echo-mini-fiio-hyby/?ref=tomcw.xyz" rel="noreferrer">Things that are offline but still digital.</a> Because we&apos;re all exhausted by being constantly connected to everything, because they sound better and because maybe, just maybe we want quality rather than convenience.</p><p><strong>Prediction 9</strong>: Someone will write a think piece about how AI will solve poverty/homelessness/inequality. The actual solution will remain &quot;give people money and stable housing&quot;.</p><p><strong>Prediction 10</strong>: Interest in Data trusts or place based cooperatives will return. At least two UK pilot projects will launch, probably in health or local government or something place based...sorry neighbourhood based. They&apos;ll struggle with governance more than technology.</p><p><strong>Prediction 11</strong>: You will wonder where Tom and his cheese toastie van is because you know it makes sense.</p><p><strong>Prediction 12</strong>: Smaller, specialised models will start outperforming general-purpose LLMs for specific social sector use cases. The &quot;foundation model&quot; hype will start to crack. Local deployment will become viable.</p><p><strong>Prediction 13</strong>: GDPR enforcement will finally catch up with AI. At least one organisation will get properly fined for AI-related data processing violations. </p><p><strong>Prediction 14</strong>: Someone will build a proper open-source alternative to Microsoft Forms that doesn&apos;t make you hate life. Please. It&apos;ll get traction in the social sector...who am I kidding, it won&apos;t, but please do it anyway.</p><p><strong>Prediction 15</strong>: The North East will take steps to establish itself as a <a href="https://tomcw.xyz/ghost/#/site/" rel="noreferrer">Public Interest AI &amp; Tech hub</a>. The pieces are there. I can hope can&apos;t I?</p><p><strong>Prediction 16</strong>: I won&apos;t do a running race, but I will run more and explore more. I will try to set <a href="https://fastestknowntime.com/route/teesdale-way-united-kingdom?ref=tomcw.xyz" rel="noreferrer">another FKT</a>, which will be probably be soundly beaten again. </p><h2 id="what-i-started-thinking-about">What I Started Thinking About</h2><p>Ok enough of that. I actually did start doing some thinking a bit more deeply about what&apos;s actually happening with AI.</p><p>Not the hype, or the the next big model, and certainly not AGI. But the gaps. The infrastructure that&apos;s missing. </p><p>I&apos;ve been thinking about what&apos;s missing in the AI ecosystem, the infrastructure gaps between &quot;look what AI can do&quot; and &quot;AI that actually serves people fairly and safely.&quot; At the moment everything seems to be about building capability or adoption, without building the infrastructure that makes capability usable, safe, accountable, fair.</p><p>We&apos;re busy pushing adoption, deploying AI agents without oversight. We&apos;re training or using models we can&apos;t edit or correct. We&apos;re fragmenting services that need connection. We&apos;re trapping people&apos;s context in silos.</p><p>And we&apos;re deploying AI without infrastructure to know if it&apos;s actually <em>helping</em>. We&apos;re training models on data scraped or bought without consent. We&apos;re optimising for efficiency without planning for inevitable failure.</p><h2 id="infrastructure-capability">Infrastructure &gt; Capability</h2><p>Here&apos;s what I think is happening: we&apos;re making the same mistake we always make with technology.</p><p>We&apos;re focused on the capability, what can it do? But are ignoring the infrastructure, the governance, how we adapt it, fail safely, learn from it, make it accountable.</p><p>I see this pattern everywhere:</p><ul><li>Build the data warehouse, ignore the data governance</li><li>Deploy the new system, skip the people bit</li><li>Launch the platform, forget the community building</li><li>Adopt the AI, miss the evaluation infrastructure</li></ul><p>I get it, Infrastructure is <em>boring</em>.</p><h2 id="what-im-actually-considering-for-2026">What I&apos;m Actually Considering for 2026</h2><p>So if I was going to build something in 2026 (and I <em>might</em>), it wouldn&apos;t be another AI model. It wouldn&apos;t be another platform. It wouldn&apos;t be another SaaS tool.</p><p>It would be infrastructure.</p><p>The boring, essential, democratically necessary infrastructure that the social sector needs if AI is going to serve people rather than extract from them.</p><p>Eight(at time of writing) specific gaps, to be precise:</p><p><strong>The Manager</strong>: Runtime supervision for AI.  Not &quot;trust the model&quot; but &quot;build the oversight layer that makes trust possible.&quot;</p><p><strong>Unlearning models</strong>: Model editing and <strong>unlearning</strong>. Because democracy runs on mutable systems, not immutable ones, and we need to be able to fix AI when we get things wrong.</p><p><strong>The Citizen AI Network</strong>: Coordination infrastructure for fragmented services. Not centralised mega-systems, but protocols that let diverse organisations and agents work together.</p><p><strong>Just In Time Interface</strong>: Component libraries for just-in-time software. Not platforms, but building blocks the sector can assemble into exactly what each organisation needs.</p><p>Edit 01/02/2026 - <a href="https://tomcw.xyz/building-blocks-for-just-in-time-software/">The Interface: Building Blocks for Just-in-Time Software</a></p><p><strong>My Memory</strong>: Citizen-owned context that&apos;s portable across services. Your data, your control, your services.</p><p>Edit 18/12/2025 - <a href="https://tomcw.xyz/my-memory-a-startup-idea-for-2026/">My Memory - A startup idea for 2026</a></p><p><strong>The Witness</strong>: Evaluation infrastructure that tracks outcomes over time. Not &quot;did the AI work?&quot; but &quot;did this help people?&quot;</p><p><strong>The Library</strong>: Community-governed training data commons. Because ethical AI needs ethical data, and ethical data needs consent, governance, and representation.</p><p><strong>The Safety Net</strong>: Fail-safe infrastructure for when AI inevitably fails. Design for resilience, not efficiency.</p><h2 id="what-happens-next">What Happens Next</h2><p>Over the next...umm few weeks(?) I&apos;m going to explore each of these eight gaps in a bit more depth. (I&apos;ll update the links as I go)</p><p>Not because I have all the answers. I pretty sure I don&apos;t, but I think these are the right questions. The infrastructure questions that get ignored while everyone chases capability.</p><p>And maybe if enough people start asking these questions, someone, maybe even me, will build the answers. </p><h2 id="one-last-prediction">One Last Prediction</h2><p><strong>Prediction 5001</strong>: In 2026, most AI conversations in the social sector will still be about capabilities and use cases. &quot;Can AI do X?&quot; &quot;Should we use AI for Y?&quot;</p><p>But a few conversations, maybe just a few, will be about infrastructure. About governance, evaluation, training data, fail-safes, coordination protocols, citizen data sovereignty.</p><p>Not the loudest. Not the most exciting. Not the ones that get all the funding.</p><p><strong>And those will be the important conversations, the ones I want to have</strong></p><p></p><h2 id></h2>]]></content:encoded></item></channel></rss>