EPISODE 1872 [0:00:11] GV: Hi, and welcome to SED News. I'm Gregor Vand. [0:00:15] SF: And I'm Sean Falconer. [0:00:15] GV: And this is the version of Software Engineering Daily where we just take a spin through the last month's headlines. We go into a main topic in the middle, where we go into a bit more depth about something very specific. And then we take a fun look at some Hacker News highlights, and then just give some predictions for the week ahead. As you may be able to hear, I'm a little bit sick this week. The news must go on, though, but I might not be as chatty as usual. Sean is going to be doing more of the chatting today. So, I thank Sean for that. [0:00:48] SF: Yeah, absolutely. I like to chat. I like to talk. So, I'm happy to step in, step up for you. But as you said, podcasting waits for no one. There's no sick days in podcasting. [0:00:58] GV: Absolutely. What have you been up to over the last month, Sean? [0:01:01] SF: Yeah. So it's been a busy month. I had some vacation time. And then the company I work for, we launched - my team launched three products recently. And now we're sort of in a sprint from now to the end of October, where we have our big annual conference, current, New Orleans this year. So, there's going to be a lot of work between now and getting ready for that, with a lot of different travel in between. It's exciting times, but lots going on. How about yourself? [0:01:27] GV: Yeah. Well, it sounds like summer is very much over for you. Yeah, it feels the same this way. I've managed to actually take a bit of a break over the last 3, 4 weeks, which is very unusual to get up in Scotland. Probably, when I come to Scotland, I always seem to get the flu or something. So, that's that. But, yeah, it's been good. But yeah, likewise, summer is definitely ending. I'm going on to some different projects in the next few weeks. So, that's exciting, but definitely ready for a lot of work ahead. With that, maybe we'll just run into the main news headlines. These are things that might have popped up on the main news outlets like Wall Street Journal, or Financial Times, or TechCrunch, whereas Hacker News is kind of more the niche stuff. So, the first one is Perplexity have been making a lot of waves. We've got a couple of things on them. [0:02:14] SF: Yeah. I mean, this is kind of a crazy one, where Perplexity offered allegedly 34.5 billion to buy Google's Chrome browser, which is almost double Perplexity's own valuation and way above what probably most people think Chrome is worth. And then while this is going on, a federal judge is weighing in antitrust remedies after ruling Google illegally monopolized search. Perplexity seems to be signaling like, "Hey, if you force a sale, we're ready to buy it." And of course, I doubt, Google's not interested in this. And they've said that selling Chrome would obviously hurt their business and even raise security risk from their perspective. Analysts don't think the sale is likely, but the bid itself shows how, I think, aggressive some of these AI companies have been. And they're very interested in getting a full hold on search. But to me, it feels like not a real acquisition attempt and more like a PR move to stay in sort of this antitrust conversation. There's a lot going on. Let's kind of attach ourselves to that. We can make some headlines. [0:03:14] GV: Yeah. Because, I mean, the bid that they made, 34.5 billion, is almost double their own valuation. Unless they're going to go get a bunch of debt on that or something. [0:03:23] SF: Yeah. Can you get a mortgage loan from a bank to buy Google Chrome? [0:03:27] GV: Yeah, exactly. We have to watch that meeting. Yeah. And as you say, Google said it's not interested, but that's the point is to get it kind of into the news around antitrust. I mean, there's been a lot of noise around Google needing to break up. I think it's funny we still talk about Google as Google, even though the parent company is Alphabet. Everybody just seems to have forgotten that. And it's still Google and all the Google products. [0:03:51] SF: Yeah. I mean, it's kind of like Facebook with Meta and stuff like that. I mean, a lot of these conversations remind me of the early 2000s with all the antitrust stuff around Microsoft as well. And there's been sort of rumors of this coming from Alphabet for quite some time. [0:04:04] GV: Yeah. But also, Perplexity, I guess maybe another PR move perhaps, but they're talking about a slightly different rev share model with publishers. [0:04:15] SF: Yeah. They want to essentially introduce this new version of their revenue share, and they plan to distribute the money when an AI assistant or search engine uses news articles to fulfill a task or answer a search question. I mean, it makes a lot of sense. If you are switching from offering blue links as search results to something where you're giving essentially a fully baked answer, can you monetize the answer in some way, where you give credit back to where the source material came from? And I think this is something we touched on last time of how might that impact the types of answers that you get. If you can essentially pay for the answer or sometimes be part of the answer, because you're paying for it, are you necessarily surfacing the best answer and doing right by the users? Now you can make the same argument for conventional search. But at least with conventional search, it's clear, well, somewhat clear, which are the ads and which are the actual organic results. I think it's something interesting to pay attention to. I am a little bit wary of, ultimately, how that could impact the value of what some of these models are able to provide users. [0:05:21] GV: Yeah. And, I mean, I think, yeah, to your point around sort of paying to play, it's going to be interesting. This is geared to publishers. Which publishers are they picking? Because if they kind of let everyone in the door, this is just going to be clickbait article heaven, where, "Oh, we're designing this article to be featured in an answer." And, interestingly, the model is moving to Comet Plus, which is their browser offering. They're offering 80% of that revenue to the publishers, which is very interesting. I mean, that's a polite way of putting it, where they're going to take literally just the revenue off a product and say, "If you're featured, we're going to distribute." I mean, there must be some kind of ratios depending on how you come up in results. But yeah, it's a very interesting where to peg it. And 80% sounds like a lot to me. But, I mean, that will come down probably. But, yeah. [0:06:16] SF: Yeah, definitely. It does seem like a lot. And I think it's one of those things where whenever there's kind of like a new form factor with which to get your message out, like a new sort of marketing channel, what inevitably happens is the early adopters of that channel will be a lot of times very successful because they're kind of like early to it. It's like if you're early to Facebook or something like that, or social media, then it's like, "Oh, wow. It's this untapped channel where everybody else is stuck sending emails. Emails is really noisy. Let's exploit this new channel." The problem is once people show success there, it's like everyone just drives in there, and eventually the channel becomes this noisy channel that gets devalued and people pay less attention to. I kind of think that we may be leaving the heyday of the chat LLM experience if this becomes too popular. [0:07:05] GV: Yeah, definitely. Moving on, next one is Intel, which has just sold 10% of itself to the US government. [0:07:12] SF: Yeah. I mean, I think this is a pretty big news in the world of semiconductors. Essentially, that 10% is worth $9 billion. It's one of the biggest government interventions in a private company since, I think, the 2008 auto bailout. The deal came directly out of negotiations between Trump and Intel's CEO Lip-Bu Tan and is framed as this way of strengthening America's chip-making leadership. And, recently, I read the government AI action plan for the US government. And part of the strategy outlined in that document was making sure that the US stays ahead of the rest of the world when it comes to CPU, GPU innovation. This seems to be somewhat in line with that. And we saw shares of Intel jump based on this news. But I think it raises some big questions of can the government ownership really turn Intel around when it comes to struggling to keep up with rivals like Nvidia? Or does this make Intel too big to fail in some fashion with taxpayers kind of now on the hook? [0:08:08] GV: Yeah, it is interesting. As you say, it's the biggest government intervention in a US company. I mean, this is a public company. There's a public shareholders as well who are looking at this and going, "Well, now, one of our shareholders is the US government." That's quite different to most companies. It does kind of have echoes of China, for example, where virtually all of their companies have state ownership in some respect. This just feels - yeah, could this be like tip of the iceberg where the government is starting to step in on certain - I mean, Intel was struggling. It won't just sort of start jumping in on things. But, yeah, I think too big to fail is a good way to think about it. Just given that these people that can fab chips effectively, it's very hard to do that. Only a couple of companies - two or maybe three can do it properly. Obviously, TSMC is the big one. Taiwanese. They're kind of in a weird gray zone between China and the US. Yeah, this just feels - I mean, I can't say I support this exactly, but I can totally understand where it's coming from. And at the end of the day, Intel's had a really rocky ride recently. They're really struggling on the CEO front. And I've still got that book, High Output Management by Andy Grove, and he was the one who took Intel to its highs, and then he died quite a little while ago. And they just don't seem to have hit their stride at all after that. I had many MacBook Pros with Intel processors that were absolutely shocking. I was very happy when Apple moved on to its own silicon. It's one to watch, certainly. [0:09:39] SF: I mean, it's a good example. I mean, who knows what's going to ultimately happen to Intel? They were one of those companies where if you went back 20 years ago, it's like hard to imagine Intel being in the state that it is now. There's all these companies that I think we talk about today, where it's impossible to think about the world without those companies, but these things happen. Not that Intel is necessarily going away, but there is always the fear of a company becoming sort of the next Xerox or something like that. And I think that some of the things that we're seeing now around this battle around search, even going back to the stuff we talked about with Perplexity, if all the high-value searches - essentially, the equivalent of high-value searches go to ChatGPT, Gemini, Perplexity, these types of interfaces, then what does that ultimately do to a behemoth in the industry like a Google? [0:10:27] GV: Yeah. Moving on to our kind of final big headline. It's Meta, unfortunately, again. But we did try not to put Meta at the top of the show again. I think they're just trying to make news. This is around the fact that they are now freezing AI hiring. [0:10:42] SF: If their strategy is to make news and press, it's a very expensive strategy. We talked a bunch about Meta AI hiring and AI talent battles that are out there in the last show. And, apparently, Meta just hit the pause on AI hiring after months-long spending spree where they poached more than 50 researchers from OpenAI, Google, Anthropic, and others. And sometimes there were offers of hundreds of millions of dollars. Zuckerberg's been personally involved in the recruiting, even paying out 14 billion for Scale AI to land as co-founders as Meta's chief AI officer. Lot of money, a lot to spend. I'm not shocked that there could be some investor pushback from that. Meta has frozen hiring, and they're reorganizing their AI division into four groups, with one focused on super intelligence. I think the other challenge, it's kind of like if you do a lot of acquisitions at once, it's one thing to be, "Hey, we acquired all these great teams, and companies, and software." It's a whole other challenge how do you actually integrate that and make sense of it? And I think they could be in a similar state where just because you acquired all this talent, it doesn't mean that talent can work together behind one singular vision and actually deliver on something. I think the jury is still out on terms of is Meta's lavish bets going to end up leading to a new generation of innovation for the company, and they'll be able to catch up in some of the AI races and offer something compelling. [0:12:06] GV: Yeah, it sort of feels like they made such a splash with the amount of money they were putting on the table or the package amounts they're putting on the table. And you can only imagine, sort of internally, there was probably just a lot of, "All right, how am I going to get on this AI team and triple my salary?" and all this kind of stuff. There was probably quite a lot going on there. But excuse the pun. I feel I did make a bit of a Meta prediction last time, which was that this could be a vanity project. I can't see how just assembling kind of the super band of AI and hoping that the album is great. I just don't know how that works. I've never seen things work where people just throw money at a bunch of desperate people, put them in a room, and say, "Go." I don't know. I can't think of many examples where that's worked. Or like go but go really fast. And we need you to go as fast as possible. That never tends to actually create great things. So, let's see. [0:13:01] SF: Yeah. And they spend a lot of money developing like the metaverse, which I don't - it wasn't wildly successful. I think if they end up in a situation where they are spending wildly on AI initiatives that don't necessarily go anywhere, it could be problematic for the company in the long run. [0:13:19] GV: Yeah, for sure. All right, so let's move on to our main topic. Something that Sean is quite the expert in. We're just talking everything agentic. We're just kind of looking at where are we with agentic, and agentic workflows, and so on. It's kind of what everyone is trying to anchor themselves to at the moment is kind of my feeling. Whether you're Anthropic or whether you're like a company that's actually been going for, say, 5 years, and now you need to tag AI on your name. But it's not just AI. You're then saying, "Oh, but we're the agentic blah-blah-blah." But the reality is kind of where we're trying to look at today. Maybe if we kind of look at this in three stages, just like where is the disconnect between reality and what's going on? What is actually happening in agentic that's real? And then just looking at any of the main innovations in agentic workflows. But I'm going to hand it over to Sean. You're the expert. [0:14:11] SF: Yeah. I mean, just, I think, looking broadly, there's a lot of, I think, recent news around what is truly the ROI for these reports on, "Hey, most of these projects are going to be cancelled in the next couple years. It's unclear what the business value is. We don't really understand the risk controls." There's all these both technical and also, I think, sort of business operational challenges around bringing these AI systems into production. And I get that. I think one of the challenges in the industry, and I often talk about this with different businesses I work with, is that there's the reality of what you can do with agentic workflows or even models in general today and where you can get value as a company. And then there's sort in the hype cycle around this, where people are vibe coding demos. Or like you see these keynotes that look really impressive and things like that. And I think there's clearly a separation between what's happening in the zeitgeist of hype and what you can do in reality. And I think, unfortunately, a lot of the hype is distracting for where there actually is value. Because I do think that there's generally value. But most of the hype is kind of focused on aspirational things that I don't think are realistic right now. [0:15:28] GV: Yeah. Like, hype being, "Oh, we can have a fully agentic employee," is essentially an extreme example. [0:15:34] SF: I think that all that does is create fear in the market, and I also think it's unrealistic. There's so many challenges right now of actually being able to build something that is fully autonomous that you could trust and run reliably. If you even think about like a multi-agent system where you have a bunch of - let's say you basically have a bunch of nodes maybe in a graph, or maybe it's a dynamic graph structure. Well, if each of those have a probability of success of, say, 90% success and 10% failure, but each of that compounds, then you end up with essentially the probability calculation, you multiply the error rate. By the end of that, if this thing's like 10 nodes deep, you might have only like a 30% chance of success. That's not a very good success rate. The idea that that could replace a person right now in most tasks is, I think, unrealistic. That being said, I think that there are a lot of sort of meat and potatoes, unsexy use cases that are tremendously useful. And a lot of it has more to do with these kind of human-on-the-loop experiences where you use a model, maybe a component of it is sort of agentic in some fashion, where there's some dynamic decision-making, and it's able to create some sort of first-pass type of document. Let's say you take support tickets and then you want to be able to give a support engineer, arm them with a potential solution or at least the materials that help them solve the problem faster. Well, I think that is a realistic scenario. And we run tests on those types of things and had humans evaluate and score those where we get like quite good success rates. And if you can make those people significantly more efficient and kind of take away some of the not-fun mundane tasks of just collecting all this data together or going inspecting certain logs, and you can have an AI system basically go and do some of that grunt work and kind of bring it together, that's very, very valuable. And I think from an ROI perspective, if you have to pay somebody to kind of run around and stitch these things together today, that's a very expensive use of human talent where I think the token cost is certainly justified. If you even think about like the legal domain, there's a whole pocket of legal where you pay people to basically go into these rooms and find a obscure document, needle in the haystack. Well, LLMs are very, very good at that kind of activity. [0:17:50] GV: Yeah. I mean, I guess, maybe two of the core examples that keep coming up as "success cases" is almost like - it's either like IT support or employee onboarding. And where does agentic become helpful in that regard? Well, with IT support, you might in Slack say, "Hey, my VPN isn't working." And it can go off and actually look up not only just the documentation for that VPN, but it can actually, for example, maybe in a browser look at the configuration of the VPN and sort of reason about why that might be a problem. The other example that again seems to be just getting reality check is good, it's coding, again. And a lot of it now is in the terminal, Claude Code, other such things. It's interesting we're already slightly moving away in the IDE to just in the terminal. What's your take on those? [0:18:46] SF: Yeah. So, I think one of the reasons why some of the stuff in engineering is kind of the tip of the spear when it comes to being able to leverage some of these, I guess, edge of what these models are capable of right now is because you have ways of kind of validating the output. Because one of the biggest challenges with a lot of these problems is it's very hard to validate whether what the model produced is actually correct or not. It might read totally fine, but it's hard to know for sure, unless you're an expert, of how do I actually validate this. In a large-scale system, how do I validate that and have some confidence at scale without having to have a human always in the loop? But with engineering type tasks, I have some checks and balances. If it's code generation, well, I can compile that code, I can run the code in a sandbox environment, I can run unit tests, I can run integration tests. I have a bunch of stuff that I can essentially have some level of confidence that it's executing correctly and doing the right thing. I think we're actually starting to see some similar things in sort of the DevOps world, where if it's something like automatically diagnosing an issue in, I don't know, let's say, cloud resources, and there are ways of essentially trying things and then undoing it. You need to be able to have something that is automated. Or at least you have like a lot of confidence around that doesn't need as much handholding. You need a way to kind of be able to roll back changes if you're allowing an agent to take certain actions. It has to be action. And then you have to be able - almost think of it like a commit log in a database, where I can run some sort of transaction, but I can undo that transaction. I have this kind of log that can go back and replay and so forth. And we've built a lot of technology around that over the years to help us solve some of these problems. And I think not all those things have been baked into some of the things that we're building and trying to productionize today. I think there's a big disconnect between being able to put together a compelling demo internally at a company and then getting to a place where you can productionize that. And a lot of those challenges have to do with how do I eval and test these? We don't have great solutions for that today. It takes a lot of hand curation to essentially do that. And how do I have some level of confidence of validating the output is correct? And then there's also this big challenge around sort of the data that describes the data. When you're building any kind of gen AI experience for a business, the big hurdle is that these models are really powerful and they're very, very smart kind of like public information, but they're kind of dumb when it comes to like your specific business, your specific task, your customer information. You have to essentially go and collect that data and provide it to the model so that you can steer it in the right direction. It has the right context, essentially, during prompt time assembly to be able to come up with an intelligent response. And there's a number of challenges there. One is simply like how do I go and retrieve the data from all these different locations? But on top of that, there's kind of this larger issue of, even if I have the data, is the model necessarily going to understand how the data relates? Because when you're talking about human language, the way these models worked and are able to get this "understanding" of human language is because they've been trained on essentially billions of documents where the relationships between words and sentences and so forth, the structure, they're able to form essentially a statistical model that helps them understand that by just seeing this multiple times. But those patterns don't exist in sort of the land of spreadsheets, and databases, and where customer data is today. The model is not going to understand that some obscurely named column in your database represents, I don't know, some sort of customer IDE that has to be factored in as a foreign key to another table without explicit instructions. This metadata, the data about the data, is really, really important. And most companies don't have a good way of essentially conveying the metadata because most of it exists inside their heads. [0:22:44] GV: I mean, have you seen any real cases? Because this feels like, from a fiction point of view, obvious. But any real cases of where companies are actually training their own foundational models on their data? Which is different to say, let's just - not putting out an abhor here, but Glean is, "Hey, we'll pull in all your data from the organization across lots of different places." And you can query that and get more contextually better results. But what we're talking about here is if you're going to set up an agentic workflow, well, which large language model are you using? And as you've just said, if you take one off the shelf, it's not going to have like this super context or understanding of your business from all sorts of angles. And, well, if it could have seen that spreadsheet before it answered the question, it would have probably given a much more accurate flow and output. Yeah, have you seen anything like that? [0:23:39] SF: I mean, some people do some level of fine-tuning. There's not too many companies I've come across that are like from the ground up building an LLM. Maybe a handful that you could count on like one hand. But most people aren't doing that. Some of them fine-tuning. But even there, there's challenges. Because with fine-tuning, since you're adjusting - not all models you can fine-tune. There's not necessarily ways of doing that if there are these API-based models that are hosted somewhere else. But then on top of that, there's a challenge of, if you are making sort of these model weight adjustments and then new models come out, or I want to try out a different model because another company comes out with the next big innovation in models, I don't have an easy way to kind of transfer that knowledge. So it can become actually a sort of barrier to innovation. And most of the people that I see doing some type of fine-tuning have very specific reasons for doing it. They're quantizing a model to shrink it down and run it on an edge device, and then they're fine-tuning it to get the performance that they need for specific tasks. Or it might be really domain-specific. But I would be cautious about doing that from the get-go. What most people are doing really is trying to figure out how do they contextualize the prompt. There's this whole area around context engineering now and also going gathering the data. And then how do you also encapsulate the metadata? There's a lot of things going on in the industry right now where I think the new battleground for data is really about the metadata. Because the most valuable, for an AI perspective, for being able to take these models and make them task specific are the most valuable data you can feed them, besides the raw data, is the data that describes the data. That's like semantic layers, knowledge graphs, ontologies. All these technologies, I think, are having a huge resurgence, but it takes a lot of work to actually get these things out of people's head and encode them so that you can make the models reliable. That's a big barrier. I think a lot of the challenges around building these systems successfully today has fundamentally to do with companies just have poor data posture to begin with. And it's all about the data. All these AI problems are fundamentally data problems. We have these powerful models, but most companies aren't in a place from a data perspective where they can even take advantage of them. [0:25:52] GV: Yeah. And that slightly runs against - at the moment, if you look at any companies selling agentic workflows, they're all saying enterprise. Well, that's where they're trying to get to. And yet, here's you saying, "Sure. But they don't have the data in a place that can make that a reality." And then, okay, well, could the enterprise do it as an internal project, for example? But it's that then classic thing of like, "Well, how much time are we going to sink into this prototype? Is it just to show everyone it's possible and then walk away from it?" Or are you, as an enterprise, really invested in trying to automate away sort of drudgery of not necessarily coding, because we're kind of doing that already, but like drudgery of other roles? Because I think the "hey, we'll automate your drudgery" is kind of a sell for most agentic to enterprise companies. But I just don't have a feeling that there's just a lot of uptake other than probably very low paid prototype type projects, basically. [0:26:52] SF: I mean, I think the things that I tend to see that actually hit production are a little bit more workflowy in nature than these kind of like truly agentic systems where you just have like a bunch of kind of nodes floating around the ether and you're going to figure out the graph dynamically. I think they're a lot more, "Hey, we want to take this particular input and be able to process it to some portion of automated processing to hand off to a person to then sort of justify." If you think about, I don't know, something loan underwriting. Well, loan could come in. That could be the input signal. There, you know, like, "Okay, all inputs are going to be a loan application." You already have constraints around what you need to interpret. It's not like anybody can just come and ask anything. It's a much more like sort of closed-world problem where I understand what the inputs are. And I can break that down into kind of like a workflow that a person might execute themselves. And there'll probably be some form of maybe dynamic decision-making as part of that. Which data resources and tools it might need to interact with? But it's a lot more fixed in terms of the graph structure of it than these truly agentic systems. And I think those are very reasonable places to start that do deliver value for business. And I see that a number of companies is being able to successfully put those into production. The other advantage of that where you kind of make it, "Hey, we're going to solve this specific problem," is, if your data is a mess, you don't have to solve all your - you don't have to boil the ocean from day one. You don't have to think like, "Oh my God. We have 150 petabytes of data. How are we going to define schemas and catalogs and have governance over all this?" Well, you don't have to do all of that. You can pick a particular thing and be like, "Okay, well, I know I need these data inputs." How do we make sure that data is a state where we can actually action it with an agent? [0:28:37] GV: And probably one area that we should touch on with, again, why do things not hit production maybe as quickly as they could? And security is kind of a big one there. Simon Willison, who has a blog that seems like half of his posts just get up to the top of Hacker News. I believe he's an exited technology person who spends most of his time writing now, which is awesome. And it's not behind a paywall. It's not behind one of those. He described this as the lethal trifecta. Well, he described this kind of before, even maybe before agentic sort of really took off. But the lethal trifecta, three things, as it would sound, access to your private data, exposure to untrusted content, and the ability to externally communicate. And this is just such a fertile place for all three of those. Access to private data. Well, companies basically only have private data. Private to themselves pretty much. Exposure to untrusted content, that's what prompts are you bringing in to drive the flow? And then what could be maliciously added in the middle there? And then the final one being then the ability to externally communicate. And we saw a case just 2 days ago, NX, which is a npm package. And that followed quite a common pattern at the moment from the externally communicate bit, which is the developer is sort of asked, "Do you want publish this into a repo?" And it does. But, A, it's a public repo. And, B, it's got a whole bunch of information in it. And that's all - you need a window of like a minute. Because whoever has figured this one out is ready to listen for that exact repo name and grab the data, and then off we go. Anyway, I've kind of gone on a little bit on that one. But yeah, what's your take on how we maybe get beyond some of these challenges? [0:30:22] SF: Yeah. I mean, it's tough. It's tough. I think that kind of going back to some of the things I said before, where what I see companies being able to successfully do are these much more sort of closed world problems. Another one of the advantages there is that it's easier to understand the security model for it because you're basically constraining the number of options. It's all about how can I add sort of as much determinism to this as possible? Because without that, where you have some chatbot-type interface with an agent who has access to all kinds of stuff, it's very hard to like police that. How do you know what someone's going to input? It's a basically unbounded problem. How do I validate whether the outputs are correct? And I think another common challenge that I see from a security perspective with companies that I talk to is that their platform teams, they want to be able to understand what each piece of software is accessing from a data perspective. And if every team is kind of running with their own independent AI projects and they're building sort of bespoke tool integrations, maybe even through MCP, which they pull down from GitHub, and they're running it themselves, or whatever, that's ridiculously difficult to control. You're sort of exponentially increasing the number of problems from where you could run into like a data security issue. You're suddenly creating this huge footprint where the platform teams, the governance teams don't really have insight into what's going on. And, ideally, the reason people invest in various - or you have a platform team and you've done your vendor procurements and you understand which data systems are there, you have some sort of central control, is to give you visibility and audit logs into what people are accessing. And I think a lot of companies are in such a hurry to rush into adopting AI that they're not always thinking about that. [0:32:10] GV: Yeah. I mean, what's maybe the TLDR here on where are we with agentic? [0:32:15] SF: I mean, my take and my advice usually for companies is try to focus on these kind of closed world problems if you want to be building stuff. There is a lot of value there. If you can take some human-driven process today that takes people considerable time to like stitch stuff together, then it's really valuable to try to arm that person with the materials for them to be successful. If you look at things even from anomaly detection to root cause analysis, that's a great place to start. There's a lot of use cases even outside of engineering where some combination of events is triggering some alarm where some human person has to go and investigate why are we having this alarm. And they have very little context other than an alarm went off. And then they have to do all this investigative work. Probably look at a bunch of different software systems just to understand why is that alarm going off, how big a problem this is. That's a great use case for having essentially, call it an agent, call it an AI system, that does a lot of that investigative work for you. You arm the person when they come in to do this work with additional materials so they can cut down the time to actually respond to that alarm and hopefully solution it. [0:33:22] GV: Awesome. Well, I think we've covered kind of the key areas on sort of where we are with how are agentic workflows actually being used at the moment I think is a bit of a confusion point because there's a lot of noise. And that's usually what we try and dispel a little bit in this main section, is take a topic where there's maybe a lot of noise and just kind of break down from a couple of angles what might actually be going on and what we've been seeing. I think it's always helpful, when you're not living in one of the sort of tech epicenters, it really is, I think, such a disadvantage. You just don't get a bunch of conversations happening around you that you might sort of suddenly understand what might be going on, actually. I hope that's been helpful for you as an audience. But, as always, we move on to our, I think, as we keep saying, kind of our favorite part of the show, where we have looked at Hacker News. We just kind of keep a running tab of a couple of things that we might want to bring up. The first one I was going to bring up - yeah, I've got one kind of it's not exactly software. I think it will relate to most people listening, which was this group has put together a proposal to ban ghost jobs. Thank you to user Tever for posting this. It got quite far up the list. And yeah, this is basically just around someone who was struggling to find a job and was getting very frustrated with thinking that actually these jobs were what are called ghost jobs, i.e., jobs that are posted by companies who have absolutely no intention to fill that role. And I've got to say, I've seen this firsthand. I know that companies do this. They put out job ads where they might still take a first interview with someone, but they have very, very little actual impetus to fill that role. And a lot of it is, yeah, often around investor sentiment. Sort of an investor maybe looks up the company, "Oh, wow. They've got 50 open positions." Yeah. And LinkedIn is quite frankly rife with these. Yeah, the group got together, and they wanted to put together kind of like a spec for what a job ad should have in it. And we've already moved to a place, I believe, in certain states in the US where you have to state the salary range. That's great. I think everyone's been appreciating that one. [0:35:28] SF: Yeah, that's the case in California. [0:35:29] GV: Right. I think New York has it, actually, as well now. These kind of two hot spots for salaries. But yeah, they're also going - some of the spec was sort of you must prove that what is the time frame that you were going to ensure that you had hired this person by, and actually state that. We need to hire this person within two 2 weeks or 2 months. You need to state whether this could have been a backfield position internally, etc. All very sensible stuff. The obvious thing is how on earth will this get implemented, actually? Who knows? I think it's great. It's not to, in any way, discourage people from taking a shot at these things. Yeah, how on earth this could get implemented? And I think it was interesting at the end of the article, there was a sort of slightly glib ending, which was, "And it seems that when people get a job, they completely forget about all about this and don't care about it anymore." Yes, it's often a problem that people searching for a job and not having a ton of luck think about more. But I do think I've at least seen it. I can't say I've experienced it firsthand on my side, but I have definitely seen basically ghost job ads go out, things that just had zero intention of being filled. It's not a fiction, is all I can say. Yeah. [0:36:39] SF: Yeah. I'm sure it happens. I think if there was something where it did become some sort of policy, a number of companies would follow it, just like stating the salary, for example. I think it would probably be difficult to police that wildly. But a lot of companies are going to comply anyway if it's sort of part of baked in the rules. Maybe that helps in some fashion. I think if you have a bunch of those sort of requirements in place, then it at least maybe makes companies think a little bit more deeply about putting the effort essentially into posting a job when they have no real intention of hiring. [0:37:12] GV: Yeah. What did you turn up, Sean? [0:37:14] SF: Yeah. I found this pretty cool blog post. I love these blog posts now where you can make them like really interactive. This post that goes over Big O notation, posted by Sam Who, is the name on Hacker News. And his website is samwho.dev. But, essentially, the explainer explains Big O notation, which is if you're not familiar, it's been a long time since you studied computer science. It's how we measure how algorithms scale as input grows. And it's essentially looks at - the article breaks down with really clear examples, constant time or big O of one, log time, linear, quadratic. And it uses these JavaScript snippets as examples with these interactive visuals to show you why, for example, looping through an array is linear, bubble sorts, quadratic, binary search is logarithmic. I've learned all those things in school and now at this point many, many years ago, but it's a really cool demonstration even if you are familiar with these concepts. And if you're just learning them, I think it's a great way to kind of reinforce what you're learning. It also calls out some coding pitfalls as well. There's some practical advice in there. If you use index of inside a loop, how does that impact performance? And what's happening sort of under the hood? You think about not just the code that you're writing, but what are maybe some of the data structures that underpin some of the libraries that you're using or some of the functions that you're using. [0:38:35] GV: Yeah, that's awesome. And as you say, yeah, this is becoming a bit more - I don't want to say popular, but seeing a few more of these sort of very interactive blog posts. I'm missing the name, but there's someone who puts out maybe once every quarter something, how bicycles work or how watches work, and it's just insane the detail and what you can do sliding things around. Yeah, I love this where it actually is coding-related and allows people to kind of maybe see some of the gotchas with stuff that, as you say, you might have learned back in high school or maybe university, but it's probably a bit of a foggy topic for you now. That's very cool. Yeah. My second one, this is one that was a bit personal to me because - admittedly, it didn't get loads of up-votes, but it crossed my radar, by user Vocrum. It was about software bug renders Wahoo GPS units unusable. Now, I'm up in Scotland, and I've got like a little head unit on my bike, and I went out and came back and found that the GPS had been pinging Iceland, and it had been pinging some other islands that were definitely not Scotland. And I was like, "Okay. Maybe it's just Wahoo having a bad day." Then I noticed that all the right data was pegged to 2006, and I was like, "Okay, well, that's like super weird." And yeah, basically Wahoo pushed out a update for their Gen 1 units. And I think it's actually just that the bug itself is kind of interesting because it's actually to do with how GPS data is transmitted. And it sounds like they had upgraded their codebase to 10-bit GPS encoding and these units just couldn't handle that. There's a whole bunch to unpack there. How did this get through? No one realized that none of these very widely sold units couldn't handle this data format. They did push out a fix within like a few days. But, yeah, basically rendered them kind of useless for like a week. I didn't also realize that timestamps appear in GPS, like transmission data. And that apparently also then affect - that was why it was pinging places like Iceland and wherever, because it was getting super confused as to like what was going on. Yeah, that was a bit of a mess. And just on a similar topic, my Chromecast had the same issue earlier this year. They pushed an update, and it just bricked the Chromecast. Now, that was really interesting because, unusually, I went and Googled it, or whatever I did, but Hacker News did it probably, before I did anything. And, actually, that was the right thing to do. Because if I had decided to turn off, turn on, that would have caused another week of fixing. Because they managed to roll out a fix for people that had not turned it off first. And then they said, "I'm really sorry. But if you've turned it off and turned it back on, we will try to fix it for you." And they did, I think, eventually. But, yeah, just this idea that we've got all these things we might have bought. I mean, certainly, the Wahoo unit is like 10 years old, but still pushing updates. But the ability to brick them from an update now is pretty high. Yeah, just watch out for that. [0:41:30] SF: The last one I had here, which is something I feel like I talked about at some point, but it's kind of interesting. Florida State University did this study where they showed that chat-based AI tools are essentially not only shaping online writing, but they're influencing everyday spoken English. They looked at 22 million words from unscripted speech, and they found a sharp increase in AI-associated buzzwords like delve. There was a period where ChatGPT overused delve all the time. I bet em- in writing is - it's less about speech, but it's being used all over the place. And, essentially, their findings suggest AI is driving this measurable language change. And I think the curious thing about this is what is sort of the widespread impact of this? We train these models basically on human language. And my thought process is like, eventually, not that too distant future, there'll be more AI-generated written content online than there is human-generated by a significant margin. And then we're training the models on that. What's that kind of do? And then on top of that, apparently, we're training our own brains and the way that we think, to like the way we're sort of influencing the way that we speak, is coming from these models. There's all this weird sort of Möbius strip bias that's coming into play here. I don't know what ends up being the large-scale impact of this, but it does worry me a little bit. [0:42:51] GV: Yeah, this is interesting, and the fact that buzzwords like delve, and intricate, and surpass were in there. I think this is interesting because Singapore, where I usually - they learn very proper English. And they say delve a lot. It's not a word that I would use very often. It's almost like, yeah, it's being trained on these very proper texts. [0:43:10] SF: That bias came from - because in Nigeria people, who speak English there in business context, they use delve more than westerners tend to use delve. And when open AI did their initial human reinforcement learning, they had hired a bunch of English speakers from Nigeria, so they ended up biasing the model during the training. They've corrected it because of people basically complained about it and stuff. But there's other biases that end up happening. You see this all the time. And that's why I think there's certain signals when you look at writing a lot of times, if you play around with these models a lot, you can tell, "Okay, this was written by AI. Or they overused AI." Because there's certain things that kind of just feel like it came from a model. [0:43:49] GV: The word I'd love to know in this context is elevate. I'm absolutely sick to death of seeing elevate in every marketing promotional material out there. But this was kind of predated AI. And my wife says this is this is AI. And I'm like, "I think elevate was just getting a bit popular. But I think AI's kind of pervaded it." I get very fed up when I see elevate. [0:44:09] SF: I mean, we already do that, where even outside of AI, if someone - a particular company is successful in marketing and uses certain terms and language, other companies copy it. It's kind of like naming of children. You can look at these spikes and name popularity and then it gets oversaturated and people stop naming their kids John for a generation. And then it comes back again because it's like what's old is new. I think it's similar with other terms as well. [0:44:38] GV: And that just gave me a funny thought, where probably in the future we'll go, "Oh yeah, there's tons of those names. That was the GPT5 era, right? Where that was the most common name that was asked for when people said, "What should I name my kid? And that name appeared with that model." And then so on, so forth. Or maybe you were named by Llama. And, again, that means you've got a slightly different name. Looking ahead, predictions as we try to just give a very completely non-serious prediction for the month ahead. What you've been thinking about, Sean? [0:45:07] SF: Yeah. I talked a little bit about this. I think that the new battle for data is going to be over metadata earlier. And also of context serving for agents and AI. And I think that market is kind of just beginning. This is not necessarily metadata, but Databricks just acquired a company called Tecton. Tecton's well known for being like a low latency feature store, but they're also now kind of have like a Gen AI offering around like context store. My prediction is essentially that. Snowflake is going to follow suit. They're going to have some sort of similar acquisition or product offering in the not too distant future. [0:45:41] GV: Gotcha. Mine is Meta will ramp up their AI hiring. No, I'm kidding. Yeah, I think mine is - well, again, this is for me just super non-serious. There will be some other device that's totally bricked by a software update. We've just had the Wahoo's Chromecast not that long ago. Who knows? Maybe like PS4s will get bricked. I've got a PS4. I've got two of them actually. It would be very fitting if those got bricked for a week or something with an update. Let's go with that. Yeah, thank you for tuning in. My apologies for how I sound. I have been enthusiastic about this episode, so don't read too much in the voice. Thank you, Sean, for really taking over on this one. I appreciate it. [0:46:24] SF: Yeah. Thank you, everyone, for tuning in. And we'll see you next month. [0:46:27] GV: See you next time. [END]