The Vibe-Coding Bubble

It's now the mid-2020s, 2025 to be specific, and the latest AI fashion is large language models (LLM) that are now so convoluted – and being fed so much data, so carefully curated, while running on vast arrays of processors that consome vast amounts of electricity and coolant water, enough to cause significant environmental harms – that they can actually do quite a good job of selecting, among things others have said, to pick which ones match a given prompt – even mixing them and matching them in ways that sometimes prove genuinely useful, if you are clever enough about constructing a good prompt – and making up really convincing bullshit (which can be fun, but gets to be a problem because folk can't see the boundary between this and the stuff parotted from others, that might have some chance of not all being bullshit).

In particular, by throwing insanely large amounts of version-control history of how code has changed (and thus, incidentally, of code) at them, it's possible to get them to help folk write code. Which is lovely and I'm happy to hear it. I'm way too close to retirement (and going off to do something completely different) for it to be worth investing significant amounts of my time in learning how to prompt them or for me to need to worry about them taking my job – indeed, if one can actually replace me, that'll be great, because my colleagues are already worrying about how they're going to do that in a few years time, when I do retire.

However, as ever in the capitalist system, there are rich people rubbing their hands with glee at the thought that they can use this to replace programmers or at least bully programmers into putting up with lower pay or worse working conditions (which would just make us less poroductive, but the idiots can't see that for all the pennies they (probably wrongly) imagine it'd save them) by claiming they could replace programmers. In this they are wildly mistaken, if only because actual programmers are – of necessity – not entirely stupid and can see that, at the very least, they'll still need someone skilled at prompting and to review the resulting slop.

An interesting experiment

Furthermore – as even one of those proponents of AI (Sam Altman, footnote 2) cited in this lovely article (by Mike Judge) said – the world wants (arguably even needs) 100 times maybe a thousand times more software, so the mere factor of ten claim he used to make for increased productivity of programmers – which the article gives compelling reasons for being skeptical about – is still way short of what we need to make up for the shortage of people with the right kind of gently demented mind to actually make good programmers.

Lest anyone suspect I might think otherwise, let me just take a moment to be quite clear that that same mind-set has its down-sides, too, and there are other mind-sets the world needs plenty of for other purposes – it's not that programmers are in any particular sense better people or anything, just one of the various types of people, for which the world currently has lots of demand. It takes, as the song says, many kinds of people to make the world go round, and the programmer type is just one of those many.

That article also includes some hard data on the author's own practial experience of how using AI on projects affected his productivity in practice. He reports (and this is a sensible measure) the ratio of how much he estimated it would take to do the job himself – estimated before he'd decided, by coin-toss, whether to vibe-code it instead – to how long it actually took, classified by which way the coin came down when he tossed it. He's done his own statistical analysis of that, and I recommend you read the article for the details, but my own eye-balling of his data says it looks like data from a negative exponential distribution (and post-hoc thinking about it does indeed have me thinking we should expect that), so the usual statistical analysis (that he most likely used, unless he knows statistics better than most; it uses the normal distribution) isn't necessarily apt. Rather, I'd say the AI increases the width scale of the distribution – and that's bad, because it not only increases the mean but also makes it harder to plan because it's noisier (i.e. it also increases the variance). Doing regular statistical analysis on the logarithms of the ratios would likely be more instructive than doing it to the raw ratios, but the general implications are going to work out similar.

Discussion

Here are some of my thoughts provoked by reading the Hacker News discussion of this:

and I think that's enough reading of the comments for me. Partly because the arguments are showing up repeatedly, but partly becaus I just know that, if I keep going long enough I'll run into side-threads so far off topic and so devoid of any semblance of sanity as to constitute a Lovecraftian horror in their own right. Internet discussion threads, how I love thee not. Besides, it's getting into the evening and I have more important things to do like reading cartoons and drinking.

How skilled geeks see LLMs

A major part of the story that tends to get missed is that most software geeks who are any damn good at their jobs would positively love to see something that deserves the name AI, precisely because it'd thin the herd – the world is so desperate for software talent that some pretty useless idiots do get hired to do technical work, only to disappoint anyone competent they might be working with. Those would be the first to go (at least from any employer worth working for) if there were a drop in demand for programmers, and better tools would enable those remaining to take up the slack in the world's continuing desperate need for significantly more than ten times as many as are presently available.

If anyone's thinking that basic economic theory claims this situation should lead any market even remotely close to free to offer big incentives to programmers, well, yes the theory does say that, and to some degree the market has done that (which has contributed to talentless fools with delusions of skilz leaping in to claim a slice of the pie and making the situation worse by being bad at it), but according to the theory the market is still out of balance, so why hasn't it corrected itself further ? Which is basicaly why a lot of competent geeks are profoundly skeptical about contemporary economic theories or claims that we live in the free markets those theories presume and (at least some of) our governments claim to provide.

The programmers who would welcome the AI revolution are typically the ones who are contradicting the current AI hype about LLMs, because we're engineers who care about what a thing actually does – not about how cool it feels to use it or how pretty or shiny it is – and understand the significance of short-comings others might dismiss as unimportant or just plain ignore. We know how easy it is to fool yourself, which is why – as Mike Judge did and reported in his article – we actually test our beliefs to find out whether we're fooling ourselves, and change our minds (again, as Mike did) when we discover we have been. This Is Just Science.

We deny that LLMs are as awesome as claimed – while being very clear about what they are good at and for, much of which really is useful and good, which many of my peers are using it for, and we're happy to have it and acknowledge some value in LLMs for it, but it's not what the hype claims – and we do so not out of sour grapes or fear that we'll lose our jobs but because it's being oversold and that's causing various kinds of harm, that are only going to stop when folk wake up and recognise its limitations, use it for what it's actually good for (there's plenty of it) and stop fooling themselves that it's good for more than it is. Wishful thinking leads to disasters.

Final stray thought

I'm left wondering whether anyone has done similar experiments for caffeine use. I realise the claims in favour of that are nowhere near as heavily hyped, but it might be interesting to see some hard experimental data.


Valid CSSValid HTML 5 Written by Eddy.