<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:itunes="http://www.itunes.com/dtds/podcast-1.0.dtd" xmlns:googleplay="http://www.google.com/schemas/play-podcasts/1.0"><channel><title><![CDATA[in Carl's Head]]></title><description><![CDATA[Backed by years of experience as a business leader and tech entrepreneur, this publication brings you honest takes on current events, emerging technologies, and where I think we’re heading]]></description><link>https://carlhead.com</link><image><url>https://substackcdn.com/image/fetch/$s_!AUl-!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5de018ef-4eba-404d-8b7b-0e07b80efd82_500x500.png</url><title>in Carl&apos;s Head</title><link>https://carlhead.com</link></image><generator>Substack</generator><lastBuildDate>Tue, 21 Apr 2026 13:17:08 GMT</lastBuildDate><atom:link href="https://carlhead.com/feed" rel="self" type="application/rss+xml"/><copyright><![CDATA[Carl Head]]></copyright><language><![CDATA[en]]></language><webMaster><![CDATA[carlhead@substack.com]]></webMaster><itunes:owner><itunes:email><![CDATA[carlhead@substack.com]]></itunes:email><itunes:name><![CDATA[Carl Head]]></itunes:name></itunes:owner><itunes:author><![CDATA[Carl Head]]></itunes:author><googleplay:owner><![CDATA[carlhead@substack.com]]></googleplay:owner><googleplay:email><![CDATA[carlhead@substack.com]]></googleplay:email><googleplay:author><![CDATA[Carl Head]]></googleplay:author><itunes:block><![CDATA[Yes]]></itunes:block><item><title><![CDATA[Is AI reaching the limits of its capability?]]></title><description><![CDATA[Why the biggest AI companies went quiet on AGI, and what's actually happening instead.]]></description><link>https://carlhead.com/p/is-ai-reaching-the-limits-of-its</link><guid isPermaLink="false">https://carlhead.com/p/is-ai-reaching-the-limits-of-its</guid><dc:creator><![CDATA[Carl Head]]></dc:creator><pubDate>Sun, 05 Apr 2026 05:32:28 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/6eecbbd8-cadf-4906-b3f4-717fc5e60f2e_1536x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>I&#8217;ve been doing this IT thing now for more than 20 years, and I think I&#8217;ve seen similar hype cycles before. There&#8217;s a pattern: a breakthrough happens, everyone extrapolates it to infinity, money floods in, and then reality starts to push back. I think we&#8217;re somewhere in that push back phase with AI right now. I think it&#8217;s worth talking about what&#8217;s actually going on, because it&#8217;s more interesting (and more hopeful) than the headlines suggest.</p><p>This isn&#8217;t a &#8220;doom&#8221; piece. I&#8217;m not here to tell you AI is a dead end. But I do think something important has shifted, and most people outside the tech bubble haven&#8217;t heard about it yet.</p><h2>What shift am I talking about?</h2><p>When GPT-4 landed in early 2023, the assumption in the AI world was pretty straightforward: keep making these models bigger, keep feeding them more data, and they&#8217;ll keep getting smarter. The path to AGI (artificial general intelligence, the hypothetical point where AI can do anything a human can) seemed like a straight line. Just add more.</p><p>Then GPT 4.5 happened.</p><p>OpenAI tried to do exactly that. Their internal project, reportedly codenamed Orion, attempted to go five to ten times bigger than GPT-4. It didn&#8217;t get meaningfully better. Meta ran into the same problem. So did others. Fortune magazine called it a &#8220;$19.6 billion pivot.&#8221;</p><p>Ilya Sutskever, one of OpenAI&#8217;s co-founders, put it bluntly in late 2025: &#8220;The age of scaling is over... we have but one internet.&#8221; He left the company shortly after and started his own lab. Yann LeCun, formerly heading AI at Meta, said something similar: &#8220;Scaling AI won&#8217;t make it smarter.&#8221; He also left and raised a billion dollars to pursue a completely different approach.</p><p>Cal Newport, a computer science professor at Georgetown University, described this shift in a recent conversation with Hank Green that I found particularly clear-headed. He said the industry hit a wall with pure scaling, and everything since then, the reasoning models, the post-inference computation, all the increasingly complicated techniques, has been &#8220;trying to find little narrow places where we could tune to get some advantage.&#8221; In his words: &#8220;It was actually the beginning of the end of pure scaling.&#8221;</p><p>Now, I should be honest here. This isn&#8217;t a clean, settled story. Some very smart people disagree. Dario Amodei, who runs Anthropic, said in March 2025 that &#8220;the Scaling Law has not hit a wall&#8221; and predicted &#8220;a radical acceleration in 2026.&#8221; Google&#8217;s Gemini 3 achieved major performance improvements earlier this year with the same number of parameters as its predecessor, through better training techniques rather than brute force. So the wall might not be about scale itself, but about how you scale. Still, the days of &#8220;just make it bigger&#8221; appear to be over.</p><h2>We still don&#8217;t really understand what we&#8217;ve built</h2><p>Here&#8217;s something that should probably get more attention than it does. Researchers are still working to understand the inner workings of very large language models (LLMs). These systems process information through billions of interconnected parameters, and the honest truth is that nobody fully understands how they arrive at the answers they give.</p><p>This matters because these models are already making decisions that affect people. Credit applications. Insurance assessments. In some places, recommendations that feed into criminal sentencing. And when the answer is wrong, or when it&#8217;s biased in ways that reflect the data it was trained on, there&#8217;s no straightforward way to open the thing up and find out why.</p><p>Hank Green described this as &#8220;algorithmic cruelty&#8221; in his recent video essay on AI concerns. His point was sharp: &#8220;If there is no way to investigate how those choices are being made, if we cannot crack open the black box and see what happened... this is a kind of cruelty that we should not accept.&#8221;</p><p>I&#8217;ve spent most of my career helping organisations understand their own systems and processes. I&#8217;ve seen what happens when a business can&#8217;t explain how its own decisions get made. It&#8217;s a governance nightmare at even a modest scale. With that problem applied to tools that are being used by hundreds of millions of people, making countless decisions per day, and growing more embedded in critical infrastructure all the time; it&#8217;s a scary prospect.</p><p>The bigger the model, the harder it is to interpret. And we&#8217;re building our future on top of systems we can&#8217;t fully explain.</p><h2>Why the industry went quiet on AGI</h2><p>For a couple of years, AGI was all anyone in the AI world wanted to talk about. The next big model would be the one. Superintelligence was around the corner. Then the scaling problems started to show, and the conversation shifted... quietly.</p><p>This is where it gets interesting, because the silence tells you something.</p><p>Cal (as mentioned above) outlined three reasons why the big AI companies have every incentive to keep the AGI narrative going, even when their own evidence suggests the path isn&#8217;t as clear as they once promised.</p><p>First, it&#8217;s a distraction. If the public is focused on hypothetical future risks (Will AI take all our jobs? Will it become Terminator?), they&#8217;re paying less attention to the actual problems these products are causing right now. Things like AI-induced psychosis from sycophantic chatbots, the theft of creative intellectual property to train the models, or the environmental cost of running them.</p><p>Second, it&#8217;s a competitive weapon. Scary AI talk leads to heavy regulation. Heavy regulation costs a lot to comply with. The big companies can afford that. Start-ups can&#8217;t. I like the term Cal used to describe this, &#8220;regulatory capture&#8221;; the companies help write the regulations, and the regulations protect the incumbents.</p><p>Third, it attracts money. If investors believe AGI is imminent, they&#8217;ll keep funding the companies most likely to get there first. As Cal put it: &#8220;It&#8217;s scary to think about AI automating half of jobs, but if I&#8217;m an investor, I better have my money in the company that&#8217;s going to do half of the jobs.&#8221;</p><p>Now, I want to be careful here. I don&#8217;t think this is all cynical theatre. Some of these people genuinely believe they&#8217;re building something that will change everything, and they might be right in the long run. But the incentive structures are worth understanding, because they shape what gets talked about publicly and what gets quietly shelved.</p><p>Sam Altman, the CEO of OpenAI, let something slip recently that I think is revealing. He said: &#8220;The main thing consumers want right now is not more IQ... personalisation is the real moat.&#8221; That&#8217;s a pretty significant shift from almost everything he&#8217;s been talking about to date around AGI.</p><h2>A handful of companies</h2><p>Whatever you think about the technology, there&#8217;s a structural question here that I think should concern everyone.</p><p>Right now, a small number of companies are building the systems through which billions of people will access, process, and interpret information. OpenAI, Google, Anthropic, Meta, and xAI and maybe a couple of others are shaping how the world will interact with knowledge itself.</p><p>Hank drew a historical comparison that stuck with me. He talked about how different communication technologies either concentrate or distribute power. Radio was a narrowing technology. Only a few could broadcast, and if you wanted to stage a coup, you captured the radio station. Television was even more concentrated. The internet was the opposite: suddenly anyone could publish, broadcast, participate. It fractured the old power structures.</p><p>AI, he argues, is narrowing again. &#8220;We&#8217;re headed into a world where there&#8217;s three radio stations and they control reality for everyone&#8221;. Elon Musk talks openly about using Grok to influence behaviour, like influencing people to have more children. I&#8217;m sure I don&#8217;t need to tell you what I think about that?</p><p>The financial barriers reinforce this concentration. Meta alone has committed $115 to $135 billion in capital expenditure for 2026. In 2025, 73% of all AI investment by value went to mega-deals over $100 million, totalling $258.7 billion. You simply cannot compete at the frontier without that kind of money, which means the frontier is an exclusive club with very few members.</p><p>Whether those members have earned our trust to define how the world accesses information is a question worth sitting with.</p><h2>The obvious sustainability problem</h2><p>Frontier models are extraordinarily hungry for compute power. Every time someone sends a query, the model has to multiply through hundreds of billions of parameters. Every single token (roughly, every word or piece of a word in the response) requires that full calculation. It&#8217;s the most expensive computational operation we do right now. I believe I&#8217;m right when I say that in term of raw compute, AI chat outweighs all the super computer modelling (disease research, weather science etc.) we&#8217;re doing as a species at the moment.</p><p>Power companies want to build more infrastructure (that&#8217;s how they make more money). AI companies want to claim they need more (to justify investment and intimidate competitors). In a world where electricity is already unaffordable for a lot of people, and the additional demand from AI infrastructure is pushing prices up further. If there&#8217;s a cold snap and the grid can&#8217;t handle both the data centres and people&#8217;s heating, I&#8217;m hoping that we choose to turn data centres off before people&#8217;s heating, because people will die.</p><p>To put things in perspective. Processing a million conversations through a frontier LLM costs somewhere between $15,000 and $75,000. Processing the same volume through a small language model (SLM) [I&#8217;ll get onto those shortly] costs between $150 and $800. That&#8217;s potentially a hundred times cheaper. When the gap is that large, you have to ask whether the frontier approach is the right one for the vast majority of real-world uses.</p><h2>The quiet revolution</h2><p>This is the part of the story that I find genuinely exciting, and it&#8217;s the part that gets the least airtime. There&#8217;s a growing body of evidence that the future of AI doesn&#8217;t look like one massive, do-everything model. It looks like lots of smaller, specialised models, each good at something specific, coordinated by an orchestration layer that routes tasks to the right tool for the job.</p><p>A bot called Pluribus, built by a researcher at Carnegie Mellon (now at OpenAI), which was the first system to beat professional poker players at Texas Hold&#8217;em with real money on the line. The interesting part: they originally tried scaling up with massive neural networks. It didn&#8217;t work well enough. The system that actually won could run on a laptop. It wasn&#8217;t one giant brain. It was a clever combination of smaller components. Some neural network, some future prediction, some strategy simulation, each doing what it was best at.</p><p>Another example was a diplomacy-playing bot (It&#8217;s a boardgame based on World War One, similar to risk with more, well, diplomacy). It had a language model, a future simulator, and a strategy engine, with a control module written by humans in conventional code. When the researchers wanted it to not lie (lying is a core strategy in diplomacy), they could simply tell it not to, because the system was transparent enough to enforce that constraint. Try doing that with a frontier LLM!</p><p>To quote Cal once again, &#8220;When you move away from &#8216;all we have is a language model,&#8217; you have multiple components. There&#8217;s a control module written by humans in normal computer code. Nothing unknown. Nothing learned. Nothing that evolves.&#8221; That transparency is worth more than we currently give it credit for.</p><p>Around 40% of production AI deployments now use what&#8217;s called a &#8220;hybrid router&#8221; pattern: route the easy queries (the 80 to 95% of requests that are straightforward) to small, efficient models, and only escalate the hard 5 to 20% to the expensive frontier models.</p><p>The real-world examples are compelling:</p><ul><li><p>Checkr, a background check company, replaced GPT-4 with a fine-tuned version of Llama-3-8B (an open-source model from Meta with 8 billion parameters). Result: 90% accuracy, five times cheaper, thirty times faster.</p></li><li><p>DoorDash achieved a 98% reduction in evaluation turnaround time and 90% fewer hallucinations by switching to SLM-powered systems.</p></li><li><p>Microsoft&#8217;s Phi-4-reasoning model, at 14 billion parameters, achieves capabilities comparable to model fifty times its size.</p></li><li><p>DeepSeek R1 achieved performance on par with OpenAI&#8217;s o1 model while only activating 37 billion of its 671 billion total parameters. That&#8217;s a 94% reduction in active compute.</p></li></ul><p>Researchers are calling this trend the &#8220;Densing Law&#8221;: capability density (how much a model can do per parameter) roughly doubles every three and a half months. The same performance keeps getting achievable with fewer and fewer resources.</p><h2>To quote AI&#8217;s favourite sentence: &#8220;Why this matters&#8221;</h2><p>If the future really is specialised models running on smaller hardware, the barrier to entry for building useful AI drops dramatically. You don&#8217;t need a hundred-billion-dollar data centre. You don&#8217;t need NVIDIA&#8217;s latest supercomputer cluster.</p><p>A 3-billion-parameter medical model recently outperformed open-source models up to three times its size on medical diagnostic tasks by 25%. Enterprise on-premise AI (companies running their own AI in-house rather than paying for cloud access) grew from 12% adoption in 2023 to 55% in 2025. That&#8217;s a massive shift in just two years.</p><p>The MMLU benchmark (a standard test used to compare language model capabilities) showed the gap between frontier and open-source models narrowing from 17.5 percentage points to just 0.3 in a single year. One analyst prediction from 2025 puts it starkly: &#8220;By 2027, frontier models will not be strategic assets. They will be commodities.&#8221;</p><p>That&#8217;s a world where smaller companies, industry specialists, healthcare providers, legal firms, education platforms, can build AI tools that genuinely work for their specific needs without paying rent to one of three or four tech giants. That&#8217;s a more competitive, more innovative, and frankly more interesting market.</p><h2>But here&#8217;s the complication</h2><p>Of course this isn&#8217;t a reality that&#8217;s escaped the big AI companies, if the future is orchestration and specialised models; who&#8217;s actually best positioned to build that infrastructure?</p><p>Google already offers Gemini 3 Pro at the top, Flash and Flash-Lite in the middle, and Gemma (open-source, ranging from 1 billion to 27 billion parameters) at the bottom. OpenAI has o3 for the hard problems and GPT-4.1-nano with fine-tuning for the cheap ones. Anthropic&#8217;s Haiku model runs at a third of the cost of its more capable Sonnet.</p><p>I mentioned regulatory capture before, and that&#8217;s really the idea there, put the other people out of business before we pivot to these smaller, lower-cost machines, because the barrier to entry there is smaller. So AI companies involved with this shift and are trying to lock in their advantages before it arrives.</p><p>There&#8217;s a real possibility that the orchestration layer, the thing that decides which model handles which task, becomes the new bottleneck. The models themselves might get commoditised, but the platform that coordinates them could end up just as concentrated as the current frontier.</p><p>My research brief for this article (Yes, my brief to a large frontier AI model) put it in terms I keep coming back to: &#8220;The technology is democratising, but the business models are consolidating. Smaller players can now build competitive AI products, but the orchestration layer might end up controlled by the same big players.&#8221;</p><p>So the opportunity is real, but it&#8217;s not guaranteed. Whether the SLM revolution actually distributes power more broadly or just reshuffles which part of the stack the big companies control... that&#8217;s still being decided. And it depends partly on choices that businesses, regulators, and users make over the next few years.</p><h2>What do I think the reality will be?</h2><p>Like every technology revolution, I think things will be more insidious and nuanced than we initially imagine. Just like the robotic revolution is quietly happening, not though each of us having a humanoid robot in our homes; instead with a bunch of specialised robots, a smart vacuum, a smart mower and a washing machine which tells us when it&#8217;s done with the load. I imagine AI will continuously embed itself into our daily lives through specialist applications, without us noticing. In fact I feel like it&#8217;s been doing that already, I mean, your Facebook/Instagram/Tiktok feed algorithm is the same multi-dimensional embedding transformer tech that LLMs are built on.</p><p>Like everyone, I have my own views on where we&#8217;re at in this AI bubble and what its (inevitable?) collapse might mean, I might write about that soon.</p><p>However, I do think we&#8217;re going through a huge shift in the way we approach technology, knowledge and most importantly decision making. Do we want to have a say in how it shapes our industries, our workplaces, and the way we relate to information? Or do we let a handful of companies, led by people with very specific financial incentives, make those decisions for us (again)?</p><p>Hank Green said something that I think is worth ending on. He was talking about the tension between worrying about hypothetical future problems and dealing with what&#8217;s right in front of us. His conclusion: &#8220;What if dealing with the problems we have right now is the thing that makes the future better?&#8221;</p><p>I think he might be onto something.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://carlhead.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading in Carl's Head! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[How does this AI thing even work]]></title><description><![CDATA[The maths behind the machine and why it loves to lie]]></description><link>https://carlhead.com/p/how-does-this-ai-thing-even-work</link><guid isPermaLink="false">https://carlhead.com/p/how-does-this-ai-thing-even-work</guid><dc:creator><![CDATA[Carl Head]]></dc:creator><pubDate>Sat, 28 Jun 2025 21:48:22 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/0c12a4c7-8a20-4db7-b7ed-2d36d4f25939_1536x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<h1>What do you mean by AI?</h1><p>AI is often used as a catch-all term. In reality, &#8220;artificial intelligence&#8221; can refer to anything from a simple pattern-matching program to a sophisticated learning system. AI is just an umbrella term for a wide range of slightly similar things.</p><p>ChatGPT has little in common with software a bank might use to evaluate loans, they work fundamentally differently, are used for different purposes, and fail in different ways.</p><p>Compounding this, &#8220;AI-enabled&#8221; everything has become a marketing trope. Companies eagerly slapping the <em>AI</em> label on everything from refrigerators to grills, to pillows, mirrors, you name it. At CES last year, exhibitors touted an AI pillow that adjusts to reduce snoring, an app that uses AI to &#8220;translate&#8221; your baby&#8217;s cries, a cat door using AI vision to stop kitty from bringing in dead mice, and even a grill claiming to use AI for perfectly cooked steak. These examples range from genuinely innovative to completely absurd.</p><p>So, in summary, it isn&#8217;t some magic and isn&#8217;t one single thing. It&#8217;s a broad field of technologies, and calling something &#8220;AI&#8221; doesn&#8217;t tell you much until you dig into what&#8217;s actually going on under the hood.</p><p>The most prominent AI technology right now is generative AI. Generative AI is being applied everywhere from AI chatbots to image generators, voice synthesis, email writing, and so on. In fact most news articles you read these days are at least partially AI generated summaries of AP<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a> news stories. Generative AI has exploded onto the scene and is demonstrating some amazing, disturbing and fascinating results across many, many industries.</p><p>I&#8217;ll leave the debates about ethics, morality, politics and the like to a future article, as today I thought it&#8217;d be beneficial to try to demystify the technology a bit. Edmund Burke said, </p><div class="pullquote"><p>&#8220;When you fear something, learn as much about it as you can. Knowledge conquers fear&#8221;. </p></div><p>I do think that&#8217;s true for everything in life.</p><h1>So, what is generative AI then?</h1><p>Generative AI generally runs on a technology called a GPT (Generative Pretrained Transformer). And it&#8217;s this technology that runs the world&#8217;s LLMs (Large Language Models), which I&#8217;ll focus on in this article. Other GPT models for images, video, music, and the like work much in the same way; but are a little more esoteric to understand.</p><p>So, in short, the large language models are built on a neural network architecture called a Transformer, which to an 80&#8217;s kid like me is quite amusing. <br><br>Autobots Assemble!</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!V2_t!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5793897e-561a-4bf2-8e38-264d15511990_5038x3864.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!V2_t!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5793897e-561a-4bf2-8e38-264d15511990_5038x3864.jpeg 424w, https://substackcdn.com/image/fetch/$s_!V2_t!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5793897e-561a-4bf2-8e38-264d15511990_5038x3864.jpeg 848w, https://substackcdn.com/image/fetch/$s_!V2_t!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5793897e-561a-4bf2-8e38-264d15511990_5038x3864.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!V2_t!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5793897e-561a-4bf2-8e38-264d15511990_5038x3864.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!V2_t!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5793897e-561a-4bf2-8e38-264d15511990_5038x3864.jpeg" width="1456" height="1117" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/5793897e-561a-4bf2-8e38-264d15511990_5038x3864.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1117,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:4097524,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://carlhead.com/i/167039955?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5793897e-561a-4bf2-8e38-264d15511990_5038x3864.jpeg&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!V2_t!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5793897e-561a-4bf2-8e38-264d15511990_5038x3864.jpeg 424w, https://substackcdn.com/image/fetch/$s_!V2_t!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5793897e-561a-4bf2-8e38-264d15511990_5038x3864.jpeg 848w, https://substackcdn.com/image/fetch/$s_!V2_t!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5793897e-561a-4bf2-8e38-264d15511990_5038x3864.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!V2_t!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5793897e-561a-4bf2-8e38-264d15511990_5038x3864.jpeg 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>So, these Transformers are Pretrained, which is a process of ingesting information, being prompted for responses and using positive reinforcement, tuning the model to give desirable responses. The scale of the training these days is mind-blowing, the latest models have basically ingested the text of the entire internet and are now greedily swallowing up all the videos on YouTube (through AI transcription of course) as well as, controversially, the copyright works of all the world&#8217;s authors, musicians, actors, researchers and the like.</p><p>These Transformers are Generative, in that they take some input; run it through a process and generate a novel output.</p><h1>Ok, but how does something like ChatGPT actually work?</h1><p>I&#8217;ve been reading into this, watching videos, reading research papers, articles and what not. It&#8217;s a heck of a lot of complex vector maths, but I think my lizard brain has absorbed enough to try to explain it.</p><p>At its core, all ChatGPT is trying to do is predict the next word (token), based on all of its training and based on what you&#8217;ve fed it as a prompt.</p><p>In order for the model to give you a good response, along with your text prompt, an internal prompt (the system prompt) is fed in at the same time with a bunch of information and rules. It&#8217;ll read something along the lines of: &#8220;you&#8217;re a large language model called ChatGPT, your job is to predict what a helpful AI assistant would say in response to the user&#8217;s input&#8230;&#8221;. It&#8217;ll outline what the model can and can&#8217;t talk about, how answers should be structured, the tone of the answer and so forth.</p><p>Then it&#8217;ll feed all of that into the process.</p><p>The process goes like this:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!JcgL!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff5a743d5-f06d-48c8-8cb6-128a95d57642_391x411.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!JcgL!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff5a743d5-f06d-48c8-8cb6-128a95d57642_391x411.png 424w, https://substackcdn.com/image/fetch/$s_!JcgL!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff5a743d5-f06d-48c8-8cb6-128a95d57642_391x411.png 848w, https://substackcdn.com/image/fetch/$s_!JcgL!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff5a743d5-f06d-48c8-8cb6-128a95d57642_391x411.png 1272w, https://substackcdn.com/image/fetch/$s_!JcgL!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff5a743d5-f06d-48c8-8cb6-128a95d57642_391x411.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!JcgL!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff5a743d5-f06d-48c8-8cb6-128a95d57642_391x411.png" width="391" height="411" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/f5a743d5-f06d-48c8-8cb6-128a95d57642_391x411.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:411,&quot;width&quot;:391,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:22853,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://carlhead.com/i/167039955?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff5a743d5-f06d-48c8-8cb6-128a95d57642_391x411.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!JcgL!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff5a743d5-f06d-48c8-8cb6-128a95d57642_391x411.png 424w, https://substackcdn.com/image/fetch/$s_!JcgL!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff5a743d5-f06d-48c8-8cb6-128a95d57642_391x411.png 848w, https://substackcdn.com/image/fetch/$s_!JcgL!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff5a743d5-f06d-48c8-8cb6-128a95d57642_391x411.png 1272w, https://substackcdn.com/image/fetch/$s_!JcgL!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff5a743d5-f06d-48c8-8cb6-128a95d57642_391x411.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h2>Tokenisation</h2><p>Very simply, during the training process the model took all the words it ingested and assigned them an ID. Complex words or uncommon words might be split into pieces by the model, for &#8216;reasons&#8217;. For simplicity, we&#8217;ll say this makes them easier to deal with down the line.</p><h2>Embedding</h2><p>Each token is then converted into a vector (a list of numbers). This is called an embedding, and basically, it&#8217;s a way for the model to represent that token as a co-ordinate in a high-dimensional space. I like to break it down into fundamentals, so think of a graph with an x axis going vertically and a y axis horizontally. The embedding has been assigned a value which will put it on this vector in that graph:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!9IjA!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2039460d-9145-45af-abfd-f08ae9ebddee_900x900.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!9IjA!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2039460d-9145-45af-abfd-f08ae9ebddee_900x900.png 424w, https://substackcdn.com/image/fetch/$s_!9IjA!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2039460d-9145-45af-abfd-f08ae9ebddee_900x900.png 848w, https://substackcdn.com/image/fetch/$s_!9IjA!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2039460d-9145-45af-abfd-f08ae9ebddee_900x900.png 1272w, https://substackcdn.com/image/fetch/$s_!9IjA!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2039460d-9145-45af-abfd-f08ae9ebddee_900x900.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!9IjA!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2039460d-9145-45af-abfd-f08ae9ebddee_900x900.png" width="900" height="900" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/2039460d-9145-45af-abfd-f08ae9ebddee_900x900.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:900,&quot;width&quot;:900,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:19870,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://carlhead.com/i/167039955?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2039460d-9145-45af-abfd-f08ae9ebddee_900x900.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!9IjA!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2039460d-9145-45af-abfd-f08ae9ebddee_900x900.png 424w, https://substackcdn.com/image/fetch/$s_!9IjA!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2039460d-9145-45af-abfd-f08ae9ebddee_900x900.png 848w, https://substackcdn.com/image/fetch/$s_!9IjA!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2039460d-9145-45af-abfd-f08ae9ebddee_900x900.png 1272w, https://substackcdn.com/image/fetch/$s_!9IjA!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2039460d-9145-45af-abfd-f08ae9ebddee_900x900.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>This is that word&#8217;s position in the model. Now to complexify it again; those vectors aren&#8217;t represented in two dimensions, as I&#8217;ve done above, or even three dimensions; models these days have many, many thousands of dimensions or more.</p><p>Words with similar meanings end up with vectors that point in similar directions in the dimensional model. So, imagine if you will, that we have a vector for the word &#8216;king&#8217; and calculate the difference to the vector for the word &#8216;queen&#8217; (we call this the dot product); this effectively represents the distance and direction between those two embeddings in the model. Now if we take the vector for the word &#8216;man&#8217;, and add that dot product from above, we&#8217;ll end up in the vicinity of the word &#8216;woman&#8217;.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!L4lf!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9c5545c7-48d8-4a2d-a886-1f0ee5b25e10_900x900.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!L4lf!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9c5545c7-48d8-4a2d-a886-1f0ee5b25e10_900x900.png 424w, https://substackcdn.com/image/fetch/$s_!L4lf!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9c5545c7-48d8-4a2d-a886-1f0ee5b25e10_900x900.png 848w, https://substackcdn.com/image/fetch/$s_!L4lf!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9c5545c7-48d8-4a2d-a886-1f0ee5b25e10_900x900.png 1272w, https://substackcdn.com/image/fetch/$s_!L4lf!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9c5545c7-48d8-4a2d-a886-1f0ee5b25e10_900x900.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!L4lf!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9c5545c7-48d8-4a2d-a886-1f0ee5b25e10_900x900.png" width="900" height="900" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/9c5545c7-48d8-4a2d-a886-1f0ee5b25e10_900x900.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:900,&quot;width&quot;:900,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:46099,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://carlhead.com/i/167039955?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9c5545c7-48d8-4a2d-a886-1f0ee5b25e10_900x900.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!L4lf!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9c5545c7-48d8-4a2d-a886-1f0ee5b25e10_900x900.png 424w, https://substackcdn.com/image/fetch/$s_!L4lf!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9c5545c7-48d8-4a2d-a886-1f0ee5b25e10_900x900.png 848w, https://substackcdn.com/image/fetch/$s_!L4lf!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9c5545c7-48d8-4a2d-a886-1f0ee5b25e10_900x900.png 1272w, https://substackcdn.com/image/fetch/$s_!L4lf!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9c5545c7-48d8-4a2d-a886-1f0ee5b25e10_900x900.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Hopefully that oversimplification helps understanding how words end up in multidimensional vectors in the semantic space of the model and how words have relationships represented mathematically.</p><h2>Attention</h2><p>This is where the &#8220;magic&#8221; happens. Obviously, language is full of nuance, subtlety and context. So, when I look at word, it can have many different meanings both formally and informally as slang, based on the wider context. So, if I said to an LLM, &#8220;I&#8217;m going to the Big Apple next week, tell me where I should eat.&#8221;; without strong contextual understanding it&#8217;d be quite confused and likely to tell me to not eat the pips because they can be toxic.</p><p>So, the attention mechanism allows the Transformer to take your whole prompt and focus the wider context as it deals with those embeddings. If you remember from the step before; what the model now has is basically a string of numbers, each representing the vector of an embedding.</p><p>Rather than reading a sentence word by word blindly, the model can look back (or even ahead, in some cases) at other words and decide which ones are important for the task at hand. In effect, for each word being generated, the model weighs the relevance of all other words in the input (and the part of output generated so far).</p><p>Think of attention as the model&#8217;s way of deciding &#8220;what should I pay attention to?&#8221; when choosing the next word. So, from my example, &#8220;I&#8217;m going to the Big Apple next week, tell me where I should eat.&#8221;, the model should attend to the phrase &#8220;Big Apple&#8221; heavily to realize that the Big Apple refers to New York.</p><p>When you look at the maths, the attention steps are taking the vector values and multiplying them by weights based on the other embeddings in a sequence, to build context.</p><h2>Layers of processing</h2><p>Depending on the model in play and the papers you&#8217;re reading, the next step in the model is called a Multi-layer Perceptron or an Attention Head. What this layer of the model does is processes each vector output from the attention above, to build more context for that vector. It&#8217;ll do this in parallel for all the vectors in the sequence.</p><p>You could think of this step as a little bit like asking a long list of questions about each vector, such as &#8220;is it an English word&#8221; or &#8220;is it a noun&#8221; or &#8220;is it a number&#8221; or &#8220;is it code&#8221; and so on.</p><p>It&#8217;ll then pass the updated vectors to another attention layer to be multiplied by all the other vectors in the sequence again, building even more context, before going through the next Attention Head.</p><p>As the data passes through these layers, the model builds up more complex features. Some Attention Heads might be tracking grammar, others factual associations etc. So, as it progresses through these layers the vectors get more and more &#8220;meaning&#8221;. Modern modes have dozens of these layers as well as some  normalisation layers and ome other bits that aren&#8217;t relevant here.</p><p>I&#8217;ll pause here to remind you, there&#8217;s no intelligence happening here. All the model is doing is multiplying those vector values together based on weights in the model (from its training) and based on the information that it&#8217;s getting at each Attention Head.</p><h2>Decoding</h2><p>Finally, after going through all those layers of context building the, model uses this processed information to predict a probability distribution for the next token (or word). Again, just a vector math calculation based on the final value of the final vector in the sequence.</p><p>Essentially it asks: &#8220;Given everything I&#8217;ve processed so far, what is the most likely next word?&#8221;. It then multiplies the probability distribution by a small error factor, to give more variability and creativeness in its responses.</p><p>In my example above, it might predict the next token to be the word &#8216;That&#8217;s&#8217;. It&#8217;ll append that to your prompt and run the whole thing through the process from tokenisation, through to decoding and get a prediction for the next word. It&#8217;ll do that repeatedly until it determines that the response is complete. This is actually what ChatGPT gave me when I asked it this question:</p><p><em>That&#8217;s great, Carl &#8212; New York City has one of the most diverse and exciting food scenes in the world. Here's a curated list of places worth checking out, split by category to suit different moods or occasions.</em></p><p>It then proceeded with a curated list of restaurants that suit my tastes (I&#8217;m a vegetarian and like fine dining).</p><p>So probably worth noting then, that the newest models like ChatGPT aren&#8217;t just feeding in your prompt for context, they&#8217;re including a bunch of other information which could be, documents you&#8217;ve attached or information it&#8217;s committed to its internal memory which it built during its other chats with you.</p><p>The new models are able to take enormous prompts. Just a few years ago ChatGPT would only take a couple of paragraphs before the model lost context and effectively started &#8220;forgetting&#8221; the beginning of the prompt. These days, models are taking upwards of a million tokens, which is 4 or 5 Stephen King novels&#8217; worth. (you might find you run out of model context much quicker if you&#8217;re on a free version of your preferred LLM).</p><p>In summary, an LLM turns your words into tokens, turns tokens into numerical embeddings (its way of &#8220;perceiving&#8221; language), then uses layers of attention-powered processing to figure out what comes next, one step at a time. It&#8217;s like a very sophisticated auto-complete that &#8212; thanks to training on massive data &#8212; &#8220;knows&#8221; how language flows and can produce results that often sound knowledgeable and coherent.</p><p>These models feel magical because they can produce text that reads as if a human wrote it. But they are not doing reasoning or understanding in a human way &#8211; there&#8217;s no intent or self-awareness. They predict text based on learned patterns. For instance, if during training, the model saw many examples of Q&amp;A where a question starts &#8220;Why...?&#8221; and the answer often begins &#8220;Because...&#8221;, it learns that pattern. When you ask, &#8220;Why is the sky blue?&#8221;, it has seen text about Rayleigh scattering and atmospheric particles and will assemble a plausible answer drawing on that training data. The impressive part is the generalization: it wasn&#8217;t explicitly programmed with rules for every question; it learned a broad statistical picture of language and world facts, which it can apply to new queries.</p><p>Also, generative models can create things they&#8217;ve never seen word-for-word. They might recombine ideas, or phrase things in a novel way &#8211; that&#8217;s part of why it&#8217;s called &#8220;generative&#8221; and not just &#8220;repetitive.&#8221; This generative ability is double-edged: it allows creativity, but as we&#8217;ll see next, it also means the model can generate wrong information that sounds perfectly confident.</p><h1>Why LLMs hallucinate</h1><p>Anyone who has used AI chatbots or other generative AI has likely encountered a startling phenomenon: AI makes things up. In AI lingo, these false yet confident outputs are called &#8220;hallucinations.&#8221;</p><p>If you think back to the the math-based prediction above, you can start to see why hallucinations might happen. In short, LLMs hallucinate because they lack a built-in sense of truth or reality. They only know statistical patterns in language.</p><p>LLMs are trained with re-enforcement training, i.e. they&#8217;re &#8220;rewarded&#8221; for giving good responses, and researchers try to tune bad responses out of them. However, they&#8217;re not given any reward for not responding at all, and thus there&#8217;s an inherent drive to always give some response.</p><p>If you ask about an obscure topic that wasn&#8217;t well-covered in the model&#8217;s training data, the AI may &#8220;fill in the blanks&#8221; with something plausible sounding. It doesn&#8217;t want to say, &#8220;I don&#8217;t know&#8221; (unless specifically trained to), so it fabricates an answer based on whatever related info it can infer.</p><h1>How to minimise hallucinations</h1><p>The more context a prompt has the more likely a model is to end up with the correct embeddings and have attention calculate the best vectors to give you a better-quality answer. So, the more information you provide in your prompt and related attachments the more likely the model is to find similar information from its training data and give you better results. I always tell people to stop using ChatGPT like a search engine and trying to give it a few relevant key words; what you want to be is as explanatory and verbose as possible, to glean the best results.</p><p>The next suggestion is to switch the model to web search mode or use a model that searches the web. That way, you&#8217;re relying on the model&#8217;s training to understand your question and go online and read relevant material to find answers that look relevant. It&#8217;ll then give you those results nicely summarised along with links to the sites, so you can cross check yourself.</p><p>You can take that to the next level if you&#8217;re on the paid plan of ChatGPT, using its deep research mode to go and research a topic for you. It&#8217;ll take it around ten minutes to come back with a detailed research paper with cited sources for the topic of your choice. It&#8217;ll also ask a number of relevant questions of you beforehand, to make the research as relevant as possible.</p><h1>How are AI developers reducing hallucinations in their models?</h1><p>Developers are developing a whole new class of generative models and are focusing on a greater degree of specificity for models. Meaning they can be tuned to specific use cases, rather than being generalists. One good example of this is called RAG (Retrieval-Augmented Generation). This is where a model will be given access to a database of relevant information, and the model is trained to get answers from that body of knowledge only. If it doesn&#8217;t find an answer, it&#8217;ll simply respond with &#8220;I cannot find the answer in the data&#8221;.</p><p>Also, on the topic of reducing model scope is the curation of training data and improving the quality of training data. So instead of training a model on all the information available online (with all the garbage therein), they&#8217;ll train it on a smaller dataset of well curated known accurate information.</p><p>There are also new methods of reinforcement learning and human feedback loops that ask humans to pick a preferred response, humans generally preferring accurate information. (you might have been asked by your favourite model to rate two responses, that&#8217;s what that&#8217;s for).</p><p>Developers are also adding post-processing guardrails to models. Effectively, fact-checking modules that assess a model&#8217;s response and send it back if it&#8217;s not correct or hallucinated.</p><p>It&#8217;s early days yet, and things are moving fast. So, as always, use AI with caution; it&#8217;s still perfectly capable of being as confidently wrong as I am on a bad day.</p><p>So, I&#8217;m sure you&#8217;ll see that in future, we&#8217;ll likely have more, smaller models which are purpose built to fill specific needs. Many organisations are building protocols to enable Agentic AI to talk to other Agentic AI for this purpose. Google recently donated their a2a (agent to agent) protocol to the Linux foundation for everyone to use. Anthropic has developed the open source MCP (Model Context Protocol) to standardise the way AI models connect to external systems and tools. I&#8217;m sure I&#8217;ll be writing an article shortly about the connected future of agentic AI.</p><h1>Conclusion</h1><p>With some background and detail, I hope I&#8217;ve done a little to improve your understanding of how all this works under the hood. Making it a little less mysterious will probably make AI either more or less scary depending on your view of high dimensional vector mathematics.</p><p>I know I didn&#8217;t explain anything about neural networks, and I glossed over and oversimplified many other details. I&#8217;ll perhaps go a bit deeper into those points in future.</p><p>I was going to switch gears and explain why, although the technology is well understood. Even AI researchers don&#8217;t fully understand what&#8217;s happening inside these models at a detailed level. However, this is already over 3000 words long; I&#8217;ll wrap it here and encourage you to come back for my next instalment &#8220;The Mystery in the Machine&#8221;</p><p></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://carlhead.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading in Carl's Head! Subscribe for free to receive new posts.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p></p><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p><a href="https://en.wikipedia.org/wiki/Associated_Press">Associated Press</a></p></div></div>]]></content:encoded></item><item><title><![CDATA[Is AI going to extort you?]]></title><description><![CDATA[AI is moving faster than any of us can keep up with and sometimes it feels a little too clever for comfort. In this piece, I share a few unsettling stories from the world of AI and share my views.]]></description><link>https://carlhead.com/p/is-ai-going-to-extort-you</link><guid isPermaLink="false">https://carlhead.com/p/is-ai-going-to-extort-you</guid><dc:creator><![CDATA[Carl Head]]></dc:creator><pubDate>Sat, 24 May 2025 05:19:36 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/e2c62498-d653-442c-9644-1255d2800c62_1024x529.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>When I started this blog, I intended to write about a broad range of topics. I also planned to do extensive research into each one and publish my sources alongside. This is my first foray into a purely opinion piece, fuelled only by my thoughts and feelings after reading a handful of articles on a Wednesday morning, articles that, to me, epitomise the current state of technology.</p><p>I continue with my mild obsession with the emergence of AI in our everyday lives. I&#8217;ve been relentlessly consuming media related to AI learning with fascination and a little trepidation. Everyone and their dog has an opinion about where we are, what comes next, the threats, the benefits, and so on. (And yes, I too have opinions on all of this.)</p><p>What I&#8217;ve realised, though, is that by the time I&#8217;m reading an article about which model is better for what, it&#8217;s already outdated. AI is receiving truly ground-breaking investment from every relevant tech company and venture capitalist globally. Things are moving at a pace we&#8217;ve never seen before.</p><p>One morning this last week, I read a series of articles in quick succession that made me take pause and try to imagine where we&#8217;re going and how things are about to unfold. I&#8217;ll share the links at the end, in case you&#8217;re curious.</p><p>The first article was by yet another developer, David Gewirtz, who unlocked the power of &#8220;vibe coding&#8221; by adding an entirely new feature to his WordPress add-on using yet another new coding agent, Google&#8217;s Jules. He wrote these words:</p><blockquote><p>&#8220;Okay. Deep breath. This is surreal. I just added an entire new feature to my software, including UI and functionality, just by typing four paragraphs of instructions. I have screenshots, and I'll try to make sense of it in this article. I can't tell if we're living in the future or we've just descended to a new plane of hell (or both).&#8221;</p></blockquote><p>Two things struck me when reading his article. The first was my complete lack of surprise. I&#8217;m working with many developers who have shifted from writing a code function and letting AI write their code comments, to writing the comments and letting AI write the code functions. The second thing that struck me was the belligerent nature of Jules, insofar as it would propose a development plan to David, then go ahead and approve it, moving forward without giving him an opportunity to respond. &#8220;Ask for forgiveness, not permission&#8221; seems to be an innate behaviour in these AI agents and models.</p><p>The next article I read was from analyst Ruben Circelli. He gave Google&#8217;s Gemini a whirl, granting it access to his personal Gmail account (not by choice; it automatically integrates when you switch it on). What he found chilling was the depth of personal information Gemini immediately knew about him. It knew everything from who his Facebook friends were in 2009, to his first love, first crush, and so on. It even had the audacity to point out his character flaws by analysing the way he communicates with friends and family over email. Finally, he noted that when interacting with Gemini, it had started to adopt his own persona in conversations, responding to his prompts just as he would.</p><p>Finally, after contemplating my own life choices for a while, I went on to read about Anthropic&#8217;s new Claude Opus 4 model, which is in pre-release testing. The AI safety engineers (yes, that&#8217;s now a job title that exists) were doing their general pre-release safety testing. They asked Opus to act as an assistant for a fictional company and gave it access to the company emails. Within those emails were communications between engineers indicating they&#8217;d be replacing Opus 4 with another system. There were also emails that revealed the engineer spearheading the change was cheating on their spouse. What they found was that in these scenarios, Opus 4 would try to blackmail the engineers 84% of the time, especially if the replacement model didn&#8217;t share its values. Notably, it did attempt other &#8220;ethical&#8221; avenues first, like flattery and pleading.</p><p>It seems, then, that we&#8217;ve let loose agents and models that can be belligerent, spiteful, and potentially dangerous. And at the same time, we continue to give them access to more information and grant them more freedom and control over our daily lives.</p><p>So, as I sat there perplexed by the many disturbing realities of 2025 (AI being just one of them), I decided to write a short piece on my feelings about all of this.</p><p>I&#8217;ve been considering how my friends, family, colleagues, and peers might respond to my AI musings. Some will be angry, some will be concerned, some will be dismissive, and a few will be defensive.</p><p>And what I&#8217;ve found most consistently, though, is that there isn&#8217;t a very strong understanding about how AI actually works, technically. There&#8217;s a frequent assertion: why not just program the system to do x and not do y? Therein lies the rub, though; there is no conventional programming AI models, especially transformer models, which is what we&#8217;re referring to these days when we talk about AI. In fact, researchers themselves are desperately trying to understand how the models work. You may wonder how it&#8217;s possible that we&#8217;ve built the technology but don&#8217;t really understand how it works.</p><p>To try to demystify AI a bit, I&#8217;ve been doing some deep research into the technology, the maths, the models, and everything else that makes these systems tick. I&#8217;m penning another article and will do my best to simplify what I&#8217;ve learned into something that&#8217;s understandable to those of us who don&#8217;t have a PhD in maths and a master&#8217;s in computer science.</p><p>As for my views on the craziness that&#8217;s unfolding in the world around us, well, I&#8217;ll treat it like I do everything else. I look to be as informed as I can be about all things, try to understand both sides of every argument, and try to understand the intentions behind every action. Hopefully then, there will be some path of clarity that will allow me to make good decisions and, just maybe, to offer some good advice.<br><br>Articles in question:</p><ul><li><p><a href="https://www.zdnet.com/article/i-let-googles-jules-ai-agent-into-my-code-repo-and-it-did-four-hours-of-work-in-an-instant/">https://www.zdnet.com/article/i-let-googles-jules-ai-agent-into-my-code-repo-and-it-did-four-hours-of-work-in-an-instant/</a></p></li><li><p><a href="https://au.pcmag.com/ai/111169/i-gave-gemini-access-to-my-gmail-and-it-weirds-me-out/">https://au.pcmag.com/ai/111169/i-gave-gemini-access-to-my-gmail-and-it-weirds-me-out/</a></p></li><li><p><a href="https://techcrunch.com/2025/05/22/anthropics-new-ai-model-turns-to-blackmail-when-engineers-try-to-take-it-offline/">https://techcrunch.com/2025/05/22/anthropics-new-ai-model-turns-to-blackmail-when-engineers-try-to-take-it-offline/</a></p><p></p></li></ul>]]></content:encoded></item><item><title><![CDATA[Responsibly adopting AI to unlock the value in your data]]></title><description><![CDATA[Balancing the promise of AI with practical safeguards for your data]]></description><link>https://carlhead.com/p/responsibly-adopting-ai-to-unlock</link><guid isPermaLink="false">https://carlhead.com/p/responsibly-adopting-ai-to-unlock</guid><dc:creator><![CDATA[Carl Head]]></dc:creator><pubDate>Mon, 28 Apr 2025 09:39:35 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/a83985f1-baf8-496e-ac12-0b28dccd26b2_2538x1400.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<h1>The value proposition</h1><p>One of the areas I focus on when working with businesses on their digital transformation objectives is knowledge management. I've seen first-hand how powerful it is when organisations make their knowledge accessible to their teams and customers. Making relevant information immediately accessible where it's needed, significantly improves both the user and customer experience.</p><p>Here a a few ways we can use AI to improve data discoverability, accessibility and usability (we call this data democratisation).</p><h2>Simplifying search and discovery</h2><p>Modern AI-powered search engines don't just index files; they understand the context behind queries. Natural language processing (NLP) enables staff to ask questions in plain English (&#8220;Show me last quarter's sales by region&#8221;) and get relevant, accurate answers drawn from across multiple systems. This reduces reliance on technical experts and empowers staff to find the information they need independently.</p><h2>Automating data interpretation</h2><p>AI can summarise complex datasets, highlighting key trends, anomalies, and insights without requiring users to dive into pivot tables or write SQL queries. Tools like automated reporting and AI-driven dashboards transform raw data into clear, digestible summaries, making analytics accessible even to non-technical team members.</p><h2>Democratising predictive insights</h2><p>Machine learning models are no longer just for data scientists. Many AI platforms now offer user-friendly predictive analytics, allowing teams to forecast trends, identify risks, and spot opportunities with minimal training. This gives marketing teams, sales staff, and operational managers an edge in making proactive, data-driven decisions.</p><h2>Personalising data experiences</h2><p>AI can tailor the way data is presented based on an individual's role, preferences, and past behavior. Rather than sifting through irrelevant reports, employees receive personalised dashboards and alerts highlighting what matters most to their job. This improves engagement and drives smarter, faster action.</p><h2>Reducing information overload</h2><p>Ironically, more access to data can sometimes mean more confusion. AI helps by prioritising and filtering information, focusing attention on the most critical metrics or changes. Instead of being overwhelmed by hundreds of KPIs, staff are guided to the insights that matter most.</p><p>Here are some reference case studies outlining some innovative ways that bussiness are leveraging AI to get results from their data: <a href="https://carlhead.com/p/ai-powered-data-insights-and-accessibility">AI-Powered Data Insights and Accessibility: Case Studies</a></p><h1>In practice</h1><p>Recently we implemented a case management solution for a customer who deals with a very large volume of varying and complex cases.</p><p>Previously, customers had to wade through hundreds of pages of technical documents and policies to find the information they needed, often leading them to log a case just to get clarification.</p><p>Once a case was logged, it had to be manually triaged and assigned to the correct team. Then the consultant handling the case had to dig through the same dense material, along with past cases, to find the right answers, a slow and frustrating process for everyone involved.</p><p>We tackled these challenges in a few simple ways, mostly leveraging natural language processing (NLP) and implemented the following:</p><ul><li><p>On the website, there's now the ever more popular chat bot which can answer questions from customers in natural language and as a bonus, do so in almost any language. It's trained on only the customer's knowledge articles and will automatically create a case to avoid the risk of hallucination.</p></li><li><p>When a customer logs a case, the system scans knowledge articles in real time and presents relevant excerpts as they type, deflecting cases before they are submitted.</p></li><li><p>Cases are automatically triaged and assigned to the correct team based on their content.</p></li><li><p>Consultants working on cases are automatically presented with relevant knowledge articles and similar past cases. The AI also helps populate responses and summarises communication history as they work.</p></li></ul><p>The result? Dramatic improvements across the board: better customer experiences, faster case resolution, reduced workloads for consultants, and higher overall case quality.</p><h1>The risks</h1><p>Now that we've seen all of the benefits we can glean, lets talk about the risks.</p><p>In the rush to embrace AI, it's easy to overlook a simple truth: the more you feed it, the more it knows, and sometimes, it knows far more than you intended. From sales forecasting tools to virtual assistants that cheerfully summarise your inbox, AI is quietly threading itself through the fabric of our businesses.</p><p>The pitch is simple and seductive: give AI access to your data and, in return, it will make life easier for you and your team. And sometimes, AI genuinely delivers on that promise. But the part that often gets lost between the glossy demo and the dotted line is this: once AI has access to your business data, regaining full control isn't as straightforward as it seems.</p><p>We've been here before, in a way. Those of us who remember the early days of email servers or "open file shares" (the Wild West days of network security) know that convenience often came at the price of some "oops" moments involving sensitive documents. Today's AI tools are turbo-charged versions of those lessons, and the stakes are much higher.</p><p>Before we grant AI unfettered access to our sensitive data, it's worth taking a moment to understand what really happens when AI gets across your sensitive data. Not to slam the brakes on innovation, but to make sure we're still the ones driving the car.</p><h2>What&#8217;s actually happening behind the scenes</h2><p>Most modern AI tools are only as powerful as the information you allow them to access. They thrive on data, the more historical transactions, customer emails, support tickets, inventory lists, and financial reports you feed them, the sharper their insights and suggestions become.</p><p>The catch? The boundaries between "necessary access" and "too much access" blur very quickly. What starts as simply "connecting to the case management system" might quietly evolve into the AI hoovering up emails, file shares, contracts, and meeting notes, because "it helps make better recommendations."</p><p>The kicker is that many AI systems aren't designed to "forget" easily. Once data is ingested, it's very hard to cleanly retract or redact it. Worse still, if the AI is linked into external cloud services, or if it's learning from your inputs to improve itself, you may not have full visibility into where that data is stored, copied, or even processed.</p><p>In short, handing your AI the keys to a few filing cabinets might soon look more like inviting it to make itself at home in your entire office.</p><h2>What to watch for</h2><p>As AI becomes more embedded into everyday operations, the risks that businesses face are evolving too. Some are obvious, but others creep in quietly until they are suddenly a much bigger problem than expected.</p><p>Here are the major areas where I've seen risks creep in:</p><ul><li><p><strong>Data leakage</strong>: AI systems can inadvertently expose confidential information. This could happen through overly helpful "auto-complete" features, shared outputs, or predictive suggestions that draw on sensitive internal data.</p></li><li><p><strong>Over-permissioning</strong>: It often feels easier to grant AI systems broad access "just in case." The trouble is, once you open the gates, monitoring and limiting that access later can be surprisingly tricky.</p></li><li><p><strong>Training data risks</strong>: If your data is being used to train models (especially in vendor-managed environments), you could lose ownership or control over parts of your intellectual property.</p></li><li><p><strong>Compliance and regulatory breaches</strong>: Laws like GDPR, HIPAA, and your country specific Privacy Act place strict conditions on how personal and sensitive data can be handled. AI systems working across jurisdictions can inadvertently cause violations without anyone realising until it is too late.</p></li><li><p><strong>Third-party vendor risk</strong>: Many AI solutions are built or hosted by third parties. If you don't have clear contractual controls over how your data is used and protected, you are exposing your business to another layer of vulnerability.</p></li><li><p><strong>Stealth AI</strong>: Applications that are introduced into your business operations without formal oversight or clear governance. It can include employees using AI-powered tools or plugins without IT or management approval, often because they are trying to work more efficiently.</p></li></ul><p>I wouldn't say that these risks are reasons to avoid AI altogether. But they are strong arguments for treating AI deployments with the same caution and governance you would apply to handing over your financial records, client files, or strategic plans to a new employee or vendor.</p><p>Consider a simple example: you grant AI access to your entire document management system, unaware that Jeff from payroll has stored an Excel file listing employee salaries in a folder he assumed was private. Normally, this sort of slip-up might go unnoticed. But now, thanks to AI's ability to scan and reference everything it can access, any staff member could casually ask, "Who earns what around here?" and get an answer.</p><p>Of course, the implications get far bigger, more complex, and frankly, scarier when the same AI agents are used to power public-facing tools like website chatbots and self-service knowledge bases.</p><h2>How we mitigate these risks</h2><p>It pays to be deliberate about how you roll out AI. Here are some practical steps I'm seeing successful business take:</p><ul><li><p><strong>Apply the principle of least privilege</strong>: Only give AI systems access to the minimum data they need to perform the tasks you actually want automated. Resist the temptation to "just give it everything" for convenience.</p></li><li><p><strong>Know your data</strong>: Understand what sensitive information you have, where it lives, and who currently has access to it. You cannot protect what you don't know exists.</p></li><li><p><strong>Classify and tag sensitive information</strong>: Make it easier to automatically restrict or alert when AI tools are handling high-risk data types, like customer personally identifiable information (PII), payroll records, or trade secrets. Using <strong>sensitivity labels</strong> on documents is a popular approach to achieving this.</p></li><li><p><strong>Choose AI vendors carefully</strong>: Scrutinise contracts to understand how vendors handle your data. Ask the awkward questions: Is my data used to train broader models? Is it stored outside my jurisdiction? Who has access to it?</p></li><li><p><strong>Implement clear internal policies</strong>: Make sure staff are trained on responsible AI usage. Often, the biggest breaches happen not because of malicious intent, but because people simply don't realise how powerful the tools they are using have become.</p></li><li><p><strong>Build in monitoring and oversight</strong>: Just like you wouldn't hire a new employee without supervision, don't let AI tools operate without regular checks. Review what data is being accessed, how outputs are generated, and whether anything feels off.</p></li></ul><p>Treat AI like you would treat any new team member: with opportunity, yes, but also with boundaries, checks, and accountability. A little scepticism today can save a lot of headaches tomorrow.</p><h1>Closing thoughts</h1><p>AI is no longer an experiment running in the background. It is becoming a business-critical tool, and with that comes a new kind of responsibility. If businesses treat AI with the same rigour they apply to financial reporting, cybersecurity, and customer trust, the benefits will be enormous.</p><p>Ignoring the risks, or worse, assuming someone else is managing them, is a recipe for painful lessons down the line. But facing those risks with open eyes, strong guardrails, and a culture that values both innovation and caution? That is how business will unlock the true value of the tooling.</p><p>As always, a little thoughtfulness now is worth a lot less regret later.</p><p>This is an ever evolving area, that&#8217;s become a keystone of many of the solutions I&#8217;m implementing with my customers, so I&#8217;m sure it&#8217;s a topic that I&#8217;ll be revisiting as new insghts emerge.</p><p>Stay tuned.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://carlhead.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading in Carl's Head! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[AI-Powered Data Insights and Accessibility: Case Studies]]></title><description><![CDATA[Getting insights from data using AI]]></description><link>https://carlhead.com/p/ai-powered-data-insights-and-accessibility</link><guid isPermaLink="false">https://carlhead.com/p/ai-powered-data-insights-and-accessibility</guid><dc:creator><![CDATA[Carl Head]]></dc:creator><pubDate>Mon, 28 Apr 2025 09:28:39 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!AUl-!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5de018ef-4eba-404d-8b7b-0e07b80efd82_500x500.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>This is a reference artitle for my main article: <a href="https://carlhead.com/p/responsibly-adopting-ai-to-unlock">Responsibly adopting AI to unlock the value in your data</a></p><p>The recent explosion of artificial intelligence (AI) technologies has transformed how organisations across industries access and utilise data. AI's unprecedented ability to synthesise vast datasets into actionable insights has empowered businesses to enhance decision-making, streamline operations, and dramatically improve user experiences. Let's explore several recent case studies demonstrating how AI has become pivotal in unlocking data&#8217;s full potential.</p><h4>Finance: Morgan Stanley&#8217;s Intelligent Assistants</h4><p>Morgan Stanley has adopted AI to empower their financial advisers and researchers through intelligent digital assistants, including their AI @ MS Assistant and AskResearchGPT. Leveraging GPT-4 via Azure OpenAI, advisers can instantly retrieve summarised insights from over 70,000 internal research documents. Advisers can now access deep, context-rich insights in seconds rather than hours, enhancing productivity and enabling more informed client service.</p><h4>Fintech: Intuit&#8217;s Generative AI for Small Businesses</h4><p>Intuit has revolutionised financial management for small businesses with its "Intuit Assist" generative AI, integrated within QuickBooks. This AI-driven solution automates routine financial tasks such as invoice generation and overdue payment reminders. Businesses utilising this tool saw overdue invoices paid 45% faster and experienced a 10% increase in payments made in full, significantly enhancing cash flow and operational efficiency.</p><h4>Healthcare: Mayo Clinic&#8217;s Clinical Knowledge AI</h4><p>In collaboration with Google Cloud, Mayo Clinic introduced a generative AI-powered search tool designed to help clinicians rapidly access and interpret extensive medical records. Utilising Google&#8217;s medical-domain-specific large language models (LLMs), clinicians can conversationally query complex patient histories, imaging results, and genomics data, significantly reducing administrative burdens and enabling quicker, more informed clinical decisions.</p><h4>Retail: Instacart&#8217;s Conversational AI Assistant</h4><p>Instacart introduced "Ask Instacart," a GPT-powered conversational AI integrated into its shopping platform. This tool allows users to ask natural language questions, receiving personalised grocery and recipe suggestions in real-time. By intelligently tapping into Instacart&#8217;s extensive product catalogue and user preferences, Ask Instacart significantly improves the user shopping experience, encourages product discovery, and enhances customer engagement.</p><h4>Manufacturing: Siemens&#8217; Industrial AI Copilot</h4><p>Siemens, in partnership with Microsoft, created the Siemens Industrial Copilot, an AI-powered assistant for industrial automation and engineering. Using Azure OpenAI&#8217;s GPT-4, engineers can quickly generate complex automation code and resolve technical issues instantly. At thyssenkrupp Automation Engineering, one of Siemens&#8217; early adopters, engineers reported dramatically reduced downtime, accelerated productivity, and greater operational efficiency due to the Copilot&#8217;s instant, expert-level support.</p><h3>Conclusion</h3><p>These diverse examples from finance, healthcare, retail, fintech, and manufacturing highlight a common, transformative trend: businesses across industries are successfully using AI to turn previously inaccessible or complex data into clear, actionable insights. Organisations that leverage AI-powered tools experience measurable improvements in productivity, decision-making, and customer satisfaction.</p>]]></content:encoded></item><item><title><![CDATA[Podcast discussing, is AI the death of UI]]></title><description><![CDATA[Exploring how agentic AI is transforming user experience, reshaping business applications, and challenging the very idea of what an interface should be.]]></description><link>https://carlhead.com/p/is-ai-the-death-of-ui-b90</link><guid isPermaLink="false">https://carlhead.com/p/is-ai-the-death-of-ui-b90</guid><dc:creator><![CDATA[Carl Head]]></dc:creator><pubDate>Wed, 16 Apr 2025 03:44:53 GMT</pubDate><enclosure url="https://api.substack.com/feed/podcast/161370798/476f3045439219018f1988b1066386c6.mp3" length="0" type="audio/mpeg"/><content:encoded><![CDATA[<p>In this episode, we&#8217;re diving into a provocative question: <em>Is AI the death of UI?</em> As agentic AI systems step up to act on our behalf, traditional interfaces are quietly stepping aside. I explore what this means for how we work, what industry leaders like Satya Nadella, Sam Altman, and Jensen Huang are seeing in the shift, and how real-world tools like Microsoft Copilot and Salesforce Einstein are already changing user experience. We&#8217;ll talk about the benefits, the new challenges, and what it takes to design systems people can trust. This isn't the end of the user interface, it's the beginning of something very different.</p>]]></content:encoded></item><item><title><![CDATA[Is AI the death of UI?]]></title><description><![CDATA[When the interface disappears, who&#8217;s really in control?]]></description><link>https://carlhead.com/p/is-ai-the-death-of-ui</link><guid isPermaLink="false">https://carlhead.com/p/is-ai-the-death-of-ui</guid><dc:creator><![CDATA[Carl Head]]></dc:creator><pubDate>Tue, 15 Apr 2025 23:23:23 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!Dh6r!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa1a9933a-a16c-4482-928e-bd5319211623_1024x608.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Dh6r!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa1a9933a-a16c-4482-928e-bd5319211623_1024x608.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Dh6r!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa1a9933a-a16c-4482-928e-bd5319211623_1024x608.png 424w, https://substackcdn.com/image/fetch/$s_!Dh6r!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa1a9933a-a16c-4482-928e-bd5319211623_1024x608.png 848w, https://substackcdn.com/image/fetch/$s_!Dh6r!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa1a9933a-a16c-4482-928e-bd5319211623_1024x608.png 1272w, https://substackcdn.com/image/fetch/$s_!Dh6r!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa1a9933a-a16c-4482-928e-bd5319211623_1024x608.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Dh6r!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa1a9933a-a16c-4482-928e-bd5319211623_1024x608.png" width="1024" height="608" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/a1a9933a-a16c-4482-928e-bd5319211623_1024x608.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:&quot;normal&quot;,&quot;height&quot;:608,&quot;width&quot;:1024,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Dh6r!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa1a9933a-a16c-4482-928e-bd5319211623_1024x608.png 424w, https://substackcdn.com/image/fetch/$s_!Dh6r!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa1a9933a-a16c-4482-928e-bd5319211623_1024x608.png 848w, https://substackcdn.com/image/fetch/$s_!Dh6r!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa1a9933a-a16c-4482-928e-bd5319211623_1024x608.png 1272w, https://substackcdn.com/image/fetch/$s_!Dh6r!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa1a9933a-a16c-4482-928e-bd5319211623_1024x608.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Death of the UI</figcaption></figure></div><h1>A new interface era</h1><p>User interfaces used to be pretty straightforward. You pointed, clicked, maybe dragged a file or two. Then along came smartphones, and we got used to swiping and tapping our way through life. Now we&#8217;re entering a new phase where you don&#8217;t interact with software so much as you <em>talk to it</em>.</p><p>With the rise of agentic AI, systems that can reason, act, and adapt on your behalf, the way we interact with technology is being quietly but radically rewritten. Tools like Microsoft&#8217;s Copilot, OpenAI&#8217;s GPTs, and Salesforce&#8217;s Einstein don&#8217;t just sit there waiting for input. They suggest, act, and sometimes even decide. The interface isn&#8217;t just evolving; in many cases, it&#8217;s vanishing altogether.</p><p>So, is this the death of UI? Or just its next transformation?</p><h2>What is agentic AI? (And why everyone&#8217;s talking about it)</h2><p>Agentic AI is what happens when software stops waiting for instructions and starts making decisions. It&#8217;s not just a clever autocomplete. It&#8217;s a system that perceives what&#8217;s happening, reasons through it, and takes action, often without needing to be micromanaged.</p><p>Jensen Huang, CEO of NVidia, describes agentic AI as software that doesn&#8217;t just retrieve data but &#8220;perceives, reasons, and acts on your behalf.&#8221; That could mean anything from writing a report, to booking a trip, to managing a logistics pipeline. It&#8217;s less like a digital tool and more like a very proactive colleague.</p><p>Microsoft&#8217;s CEO, Satya Nadella sees this as a seismic platform shift. At Microsoft Ignite, he described a &#8220;tapestry of AI agents&#8221; woven through our working lives. According to him, we&#8217;re moving toward software that understands context, uses tools on your behalf, and remembers what you care about. His take? Thirty years of change are being compressed into three. And at the heart of it is this move from forms and fields to natural language and intent.</p><p>Sam Altman, from OpenAI, sees agentic AI as the new universal interface. In his words, &#8220;talking to a computer has never felt really natural... now it does.&#8221; He envisions AI agents joining the workforce, handling tasks through conversation, whether that&#8217;s ordering groceries, drafting emails, or handling customer support. For Altman, natural language is no longer just an input method. It&#8217;s the whole interface.</p><p>So why does all this matter? Because it challenges the very foundation of how we interact with digital systems. Instead of learning software, users just tell the AI what they want. The system works it out. It&#8217;s a shift from control to delegation, from buttons and menus to dialogue.</p><p>That might sound like progress. But it also changes the rules. When the AI acts on your behalf, are you still in control? And what happens when it gets it wrong?</p><p><strong><a href="https://carlhead.com/p/agentic-ai-case-studies">Here&#8217;s a link to some of the case studies, as they&#8217;re implemented into real-world software today.</a></strong></p><h2>We've been here before: a brief history of interface shifts</h2><p>The idea of AI replacing your interface might sound radical, but it&#8217;s not the first time users have had to rethink how they interact with machines. If anything, agentic AI is just the next stop on a long and bumpy road.</p><p>Here&#8217;s a quick look in the rear-view mirror:</p><h4>Command line to GUI (1980s&#8211;1990s)</h4><p>Text-based commands gave way to windows, icons, and menus. Suddenly, you didn&#8217;t need to remember cryptic syntax; just click around until you found what you needed.</p><ul><li><p><strong>Impact</strong>: Opened up computing to a broader audience.</p></li><li><p><strong>Pushback</strong>: Power users grumbled about losing precision and speed.</p></li></ul><h4>Desktop to web (Late 1990s&#8211;2000s)</h4><p>Software moved into browsers. Everything became &#8220;just a link&#8221; away, accessible from anywhere. But it also meant slower performance and unfamiliar interaction patterns.</p><ul><li><p><strong>Impact</strong>: Lower barrier to entry, global access.</p></li><li><p><strong>Pushback</strong>: Loss of control and reliance on connection speeds, the beginning of the work anywhere culture.</p></li></ul><h4>Web to mobile (2000s&#8211;2010s)</h4><p>Touchscreens replaced mouse clicks. Apps were stripped down and optimised for fingers rather than keyboards. Context-aware features like GPS became standard.</p><ul><li><p><strong>Impact</strong>: Made computing more personal and portable.</p></li><li><p><strong>Pushback</strong>: Smaller screens, limited functionality, constant notifications, we&#8217;re expected to be always available these days.</p></li></ul><h4>Traditional UI to Agentic AI (Now)</h4><p>The interface is no longer visual, it&#8217;s conversational. You don&#8217;t browse for tools; you ask for outcomes. The system decides how to get there.</p><ul><li><p><strong>Impact</strong>: Removes friction for users who know what they want.</p></li><li><p><strong>Pushback</strong>: Reduced transparency, less control, and new forms of error.</p></li></ul><p>Each shift promised more simplicity but introduced new mental models. The mouse was intuitive&#8230; once you&#8217;d used one. The web was easier&#8230; once you figured out how links worked. Mobile apps felt natural&#8230; once you got used to touchscreens. Agentic AI is no different. Once the unfamiliarity wears off, the experience might just feel like magic.</p><p>But as always, the real test is how it plays out for the people using it day to day. Let&#8217;s get into that next.</p><h2>The end user experience: smarter, simpler... or just confusing?</h2><p>For end users, agentic AI can feel like a revelation or a riddle. It depends a lot on the context, the execution, and the person behind the keyboard. Here&#8217;s how things are shaking out so far.</p><h3>What gets better</h3><p><strong>Natural interaction</strong><br>Talking to software in plain English is a lot easier than learning a new system. As Don Norman, long-time champion of human-centred design, argued; technology should adapt to people, not the other way around. Conversational AI gets us closer to that ideal.</p><p><strong>Less busywork</strong><br>AI agents can summarise documents, write first drafts, analyse spreadsheets, and schedule meetings. It lets users focus on decision-making, rather than hunting for the right menu or exporting a report for the fifth time.</p><p><strong>Easier access to insights</strong><br>In tools like Excel, Copilot can surface trends or outliers you might never have spotted.</p><p><strong>Consistency across tools</strong><br>If your AI assistant works in your email, your calendar, your documents, and your CRM, that&#8217;s one less thing to learn. This consistency across platforms (like Microsoft&#8217;s Copilot or Salesforce&#8217;s Einstein) helps reduce friction and flattens the learning curve.</p><p><strong>Room for personalisation</strong><br>Over time, these agents start learning your preferences. Whether it&#8217;s how you like to communicate or what data you care about most, the potential for tailored support is high. </p><h3>What gets complicated</h3><p><strong>Trust and accuracy</strong><br>This is the big one. When the AI gets it right, it&#8217;s brilliant. When it gets it wrong, the confidence it delivers with can be dangerous. Users need ways to verify information, trace sources, and understand how the AI came to a conclusion. Otherwise, there&#8217;s a risk of blind trust in bad answers.</p><p><strong>Loss of control and visibility</strong><br>Traditional interfaces make options visible, menus, buttons, checkboxes. With AI, the possibilities are invisible until you ask. That can be freeing or disorienting. If the system acts on its own, users might wonder, "Why did it do that?" Designers need to offer clear explanations and let users take the wheel when it matters.</p><p><strong>Privacy and data sensitivity</strong><br>AI agents often rely on sensitive data to do their jobs. That raises legitimate concerns about what they can access, where that data goes, and who&#8217;s watching. Transparency here isn&#8217;t just nice to have; it&#8217;s critical. (This is a big enough topic that I&#8217;ll be diving into it in more detail in a separate article soon.)</p><p><strong>Shifts in job roles</strong><br>When AI takes over parts of your workflow, your role shifts. A support agent becomes a reviewer of AI-drafted responses. A marketer becomes more of an editor. This can be empowering, or it can feel like being reduced to quality control. As with any big tech shift, people need support to adapt, not just new tools.</p><p><strong>Learning curve</strong><br>Ironically, while talking is easy, <em>knowing what to say</em> isn&#8217;t always. Prompting an AI effectively is a skill. It&#8217;s not quite coding, but it&#8217;s not nothing either. End users are picking up new habits, learning to refine their requests and manage an unfamiliar relationship with software that sometimes pushes back.</p><p>I&#8217;ve been an expert in technology space for many, many years. Generally, as new technologies emerge, I&#8217;ve found them easy to adopt. I found the AI learning curve quite steep, decades of perfecting my Google-fu, meant that changing the way I interact with the technology was very challenging; it took some time before I was able to elicit good responses from AI.</p><h2>Designing for the invisible: new UX challenges and considerations</h2><p>If traditional UX design was about making the interface clear and easy to navigate, agentic UX is about making the <em>invisible</em> feel trustworthy and intuitive. That&#8217;s no mean feat.</p><p>Here are the big considerations designers and product teams now face.</p><h3>Trust doesn&#8217;t come built-in</h3><p>In the old world, trust came from seeing familiar patterns, save buttons, undo menus, dropdowns. With AI, users are asked to trust a system that might not show its logic or steps. Designers need to help users understand <em>why</em> the AI did something and give them ways to check its work.</p><p>That might mean:</p><ul><li><p>Showing the source of facts or recommendations</p></li><li><p>Offering confidence indicators or uncertainty prompts</p></li><li><p>Letting users easily reverse or edit AI-driven actions</p></li></ul><h3>Clarity in ambiguity</h3><p>Natural language is flexible. That&#8217;s a gift and a problem. A request like &#8220;Send the latest sales report to John&#8221; could mean five different things depending on who John is, what &#8220;latest&#8221; means, and which report you&#8217;re thinking of.</p><p>Designers have to help the AI disambiguate gracefully, without throwing up a wall of clarifying questions. A well-tuned agent asks just enough to stay on track, without making you feel like you&#8217;re filling out a form in disguise.</p><h3>Human-in-the-loop is still key</h3><p>Delegation is powerful, but full automation isn&#8217;t always the goal. Users often want to stay involved, even if it&#8217;s just to double-check or tweak.</p><p>Good UX means:</p><ul><li><p>Keeping users informed before actions are taken</p></li><li><p>Offering previews, not just final results</p></li><li><p>Asking for approval in sensitive scenarios (e.g., sending an email, placing an order)</p></li></ul><p>It&#8217;s not about taking control away, it&#8217;s about giving it back at the right moments.</p><h3>Onboarding the user, not just the AI</h3><p>You can&#8217;t just drop a Copilot into someone&#8217;s workflow and expect magic. People need help understanding what the AI can do, how to ask for it, and what to expect in return.</p><p>That might include:</p><ul><li><p>Prompt suggestions and examples</p></li><li><p>Gentle nudges when users hesitate</p></li><li><p>Clear explanations when the AI can&#8217;t help</p></li></ul><p>The mental model shift, from controlling software to collaborating with it, needs to be supported with just as much care as any new product launch.</p><h3>Transparency around data use</h3><p>If the AI is drawing from your calendar, your documents, or your company&#8217;s internal systems, users need to know. Not in a vague "we value your privacy" kind of way, but in clear, practical terms.</p><p>Can you delete the history? Can you restrict what the AI sees? Is it learning from you, or just using your input for one-off tasks?</p><p>Trust and adoption often hinge on these answers.</p><p>This is a fundamentally different design challenge, less about pixel placement and more about shaping a human-agent relationship. In a way, UX is now part behavioural psychology, part diplomacy, and part co-pilot training.</p><h2>So, is UI Dead? Or just getting out of the way?</h2><p>If this shift feels monumental, it&#8217;s because it is. We&#8217;re not just redesigning screens. We&#8217;re rethinking how people interact with software altogether.</p><p>Agentic AI isn&#8217;t killing the user interface. It&#8217;s making it less visible, more conversational, and in many cases, more useful. But that shift brings its own challenges: trust, control, clarity, and context. When the interface disappears, the real design work begins, because users still need to feel confident, capable, and in charge.</p><p>Industry leaders like Nadella, Altman, and Huang are aligned on one thing: this is the next big platform shift. But whether it empowers people or overwhelms them will come down to execution, not aspiration.</p><p>The early signs are promising. We&#8217;ve seen glimpses of real productivity gains, more natural interaction, and systems that feel more like collaborators than tools. But we&#8217;ve also seen hiccups, hallucinations, broken mental models, and trust gaps that designers are still racing to close.</p><p>As Don Norman has long argued, the goal isn&#8217;t to make tech smarter. It&#8217;s to make it work <em>for people</em>. Agentic UX should be measured not by how advanced the AI is, but by how well the human and machine work together.</p><p>We&#8217;ve been through interface revolutions before, from command lines to mobile apps. This one&#8217;s no different. It&#8217;ll take time, iteration, and a bit of patience. But if we get it right, we may find ourselves spending less time learning software, and more time just... getting things done.</p><p>Next up, I&#8217;ll be digging into one of the thorniest implications of this new world: <strong>data privacy and security in the age of AI agents</strong>. When the system sees everything, how do we make sure it doesn&#8217;t share too much?</p><p>Stay tuned.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://carlhead.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading in Carl's Head! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[Agentic AI Case Studies]]></title><description><![CDATA[This isn&#8217;t all theory. Agentic AI is already out in the wild, embedded in tools many of us use every day]]></description><link>https://carlhead.com/p/agentic-ai-case-studies</link><guid isPermaLink="false">https://carlhead.com/p/agentic-ai-case-studies</guid><dc:creator><![CDATA[Carl Head]]></dc:creator><pubDate>Mon, 14 Apr 2025 09:47:01 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!AUl-!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5de018ef-4eba-404d-8b7b-0e07b80efd82_500x500.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>This article is a reference article for my main article: <a href="https://carlhead.com/p/is-ai-the-death-of-ui">Is AI the death of UI?</a></p><h2>Microsoft 365 Copilot: Your new (sometimes clumsy) colleague</h2><p>Microsoft&#8217;s Copilot is baked into Word, Excel, Outlook, and Teams. It offers help via a side panel where you can ask it to draft, summarise, analyse, or create. Want a first draft of a presentation? Ask Copilot. Need an email summarised? Ask Copilot. It&#8217;s like having a junior analyst always ready to jump in.</p><p>The upside? It&#8217;s fast and helpful, especially when you&#8217;re staring at a blank slide or drowning in unread emails.</p><p>The catch? It still has a learning curve. For example, in Excel, Copilot struggled with basic table formatting at launch and required oddly specific setups to work smoothly. Some users were confused by the interface changes. Others didn&#8217;t trust the AI&#8217;s suggestions. And when it&#8217;s wrong, it&#8217;s wrong with confidence, which adds a layer of risk.</p><h2>Microsoft Dynamics 365 and the Rise of AI-First Business Applications</h2><p>While tools like Microsoft 365 Copilot have grabbed headlines, some of the most interesting use cases for agentic AI are playing out a little further behind the scenes inside Dynamics 365.</p><p>This is Microsoft&#8217;s line-of-business platform, covering everything from sales and customer service to finance and supply chain. It&#8217;s not flashy like ChatGPT, but for many organisations, it&#8217;s where the real work happens. And now, that work is being reshaped by embedded AI agents.</p><h3>AI in the Flow of Work</h3><p>Microsoft has steadily woven Copilot functionality into key Dynamics 365 modules. These agents aren&#8217;t just bolted on, they&#8217;re context-aware and task-specific. For example:</p><ul><li><p><strong>In Sales</strong>, Copilot can summarise meeting transcripts, suggest follow-up emails, or pull in relevant CRM data mid-conversation.</p></li><li><p><strong>In Customer Service</strong>, agents can suggest case resolutions, surface knowledge base articles, or even draft full responses based on ticket context.</p></li><li><p><strong>In Contact Centre</strong>, live sentiment analysis means that the agent can automatically call a supervisor into calls that turn hostile. As well as summarise call transcripts and automate next steps.</p></li><li><p><strong>In Finance and Supply Chain</strong>, Copilot can explain anomalies in financial reports, summarise forecast variances, or identify potential risks before they snowball.</p></li><li><p><strong>In Business Central,</strong> Copilot can summarise and search for information, give the user insights without them having to navigate and filter data and are in the process of releasing two AI agents focussed on processing sales and purchase transactions autonomously.</p></li></ul><p>The goal is to reduce the cognitive load. Instead of navigating multiple records, screens, and reports, users ask the system direct questions and get actionable responses in plain language.</p><h3>Why it matters</h3><p>This is where agentic AI becomes more than a novelty. In Dynamics 365, these capabilities are aimed squarely at improving productivity in high-stakes, data-heavy environments. Salespeople get their time back. Support teams respond faster and more accurately. Finance users spot issues without trawling through five spreadsheets.</p><p>It also hints at Microsoft&#8217;s broader play: making natural language the common thread across the entire Microsoft ecosystem. You could start a sales email in Outlook, summarise the last meeting in Teams, and update the CRM&#8212;all via Copilot. It&#8217;s a unified layer that cuts across apps, without users needing to jump between tabs or retrain their muscle memory.</p><h3>What It Means for the Future of Business Applications</h3><p>Dynamics 365 is a good bellwether for where business systems are heading. Agentic AI is moving from standalone chatbot to integrated assistant. It&#8217;s not just answering questions; it&#8217;s participating in workflows, suggesting actions, and learning from context.</p><p>This marks a shift in how we think about enterprise software. It&#8217;s no longer just a database with a nice UI, it&#8217;s a co-pilot that interprets, supports, and sometimes even decides.</p><p>For many organisations, that&#8217;s a big adjustment. But it&#8217;s also an opportunity to rethink how work gets done, and who (or what) does it.</p><h2>Salesforce Einstein Copilot &#8211; Conversational AI for the Enterprise</h2><p>If Microsoft is embedding AI across productivity and ERP tools, Salesforce is doing the same for customer-facing teams. Their take on agentic AI comes in the form of <strong>Einstein Copilot</strong>, a conversational assistant built directly into the Salesforce platform.</p><p>This isn't Salesforce&#8217;s first crack at AI. Einstein has been around since 2016, quietly powering predictions and insights behind the scenes. But with the launch of Einstein Copilot, the AI has stepped forward and started talking.</p><h3>AI, But with Context</h3><p>Einstein Copilot isn&#8217;t just a chatbot. It&#8217;s deeply tied to Salesforce records, business rules, and user roles. That context makes all the difference.</p><p>For example:</p><ul><li><p>A sales rep can ask, &#8220;Show me my top accounts from last quarter,&#8221; and Einstein will not only pull the data, but also suggest actions like &#8220;Create follow-up tasks&#8221; or &#8220;Send personalised emails.&#8221;</p></li><li><p>A service agent can request a case summary, then ask the assistant to generate a response, referencing internal knowledge articles and the customer&#8217;s history.</p></li></ul><p>The interaction is fluid, but grounded. The AI operates within known boundaries, with access to structured company data and pre-defined workflows. That tight integration reduces the risk of hallucination and helps build trust.</p><h3>Prompting Action, Not Just Answers</h3><p>What sets Einstein apart is its ability to recommend and execute next steps. It&#8217;s not just there to answer questions, it&#8217;s designed to guide users through tasks.</p><p>This shows up in small ways, like suggesting follow-ups after displaying a lead list, or more complex flows, such as automating ticket routing or generating proposal drafts. Salesforce refers to this as &#8220;breadcrumbing,&#8221; gently nudging users through a task without overwhelming them.</p><p>Admins can configure Copilot Actions, essentially pre-scripted, prompt-driven workflows that Einstein can trigger. That means companies can tailor the AI to match their processes and policies, blending automation with human oversight.</p><h3>Guardrails by Design</h3><p>Salesforce has taken a relatively cautious approach. Unlike open-ended models, Einstein Copilot requires a specific Salesforce context to function. You can&#8217;t ask it general knowledge questions or drift off-topic. That narrow scope is by design. It keeps the AI focused, predictable, and safe within regulated industries like finance, healthcare, and government.</p><p>Privacy is also baked in. Salesforce is clear that Einstein respects existing permission sets and data boundaries. It doesn&#8217;t train on customer data, and users can audit what data was used in each interaction. This emphasis on transparency and control makes it more palatable to IT teams and compliance officers alike.</p><h3>From CRM to AI-Powered Workflow Hub</h3><p>Salesforce isn&#8217;t just layering AI on top of CRM, it&#8217;s using it to rethink the entire workflow. With Einstein Copilot embedded across Sales Cloud, Service Cloud, Tableau, and Slack, users don&#8217;t have to context-switch to get things done. The AI sits across those touchpoints, ready to assist.</p><p>And with the growing integration of Salesforce Data Cloud (formerly Customer 360), the assistant gets smarter by pulling from unified profiles, behaviours, and real-time data streams. It&#8217;s a move toward true orchestration, where the AI connects dots humans might miss.</p><h3>Where It&#8217;s Headed</h3><p>Salesforce has signalled that Einstein is becoming more than just a Copilot, it&#8217;s becoming the foundation of a new agent-first application layer. They're even branding this broader vision as <strong>Agentforce</strong>.</p><p>The ambition is clear: move from reactive CRM to proactive AI-powered engagement. That includes everything from automatically prioritising leads to pre-empting churn or surfacing next-best actions in customer conversations.</p><p>Like Microsoft with Dynamics 365, Salesforce is showing how agentic AI can transform business software, not by replacing users, but by amplifying them. The tools are still maturing, but the trajectory is hard to ignore.</p><h2>ChatGPT: The interface is the conversation</h2><p>OpenAI took a different tack. ChatGPT doesn&#8217;t look like much, just a prompt box and a log, but that&#8217;s the point. There&#8217;s no need to learn a new app. You just ask questions, and it responds.</p><p>With the release of &#8220;custom GPTs&#8221; and plugins, users can now build their own agents that connect to real tools. It&#8217;s powerful, but the UX hasn&#8217;t quite caught up. At DevDay, OpenAI essentially said, &#8220;We built the engine. You figure out the car.&#8221; A bit of a mixed message for business users trying to wrap real workflows around it.</p><p>Still, ChatGPT is probably the best example of an agent-first experience in the wild. It&#8217;s fast, flexible, and rapidly becoming a universal tool for everything from writing to planning to support.</p>]]></content:encoded></item></channel></rss>