Category: Computing

When DOES it make sense to use AI?

I created my first neural network back in the late 90s, as part of my Ph.D, to do handwriting recognition on images of whiteboards. It wasn't a very good network; I had to write the whole thing from scratch as there weren't any suitable off-the-shelf libraries available, I didn't know much about them, and I didn't have nearly enough training data. I quickly abandoned it for a more hand-tailored system. But one of the early textbooks I was reading at the time had a quote, I think from John S Denker, which I've never forgotten: "Neural networks are the second-best way to do almost anything."

In other words, if you know how to do it properly, for example by evaluating rules, or by rigorous statistical analysis, don't try using a neural network. It will introduce inaccuracies, unpredictability, and make it very much harder either to prove that your system works, or to debug it when anything goes wrong.

The problem is that there are many situations in which we don't know how to do it 'properly', or where writing the necessary rules would take far too much time. And 'machine learning', the more generic term encompassing neural networks and similar trainable systems, has advanced amazingly since I was playing with it. For many tasks, we also now have masses of data available, thanks to the internet. (I was playing with my toy system at about the same time as I was experimenting with these brand new 'web browsers'.) So while it remains the case, as a Professor of Computer Science friend of mine likes to put it, that "Machine learning is statistics done badly", it can still be exceedingly useful. It would almost certainly be the right way for me to do my handwriting-recognition system now, for example, and over the last few decades we've discovered lots of other pattern-matching operations for which it is essential - analysing X-rays for evidence of tumours is just one example where it has saved countless lives.

But all of this is nothing new. So why the current excitement about 'AI'? After all, 'artificial intelligence', like 'expert system', is one of those phrases we heard a lot in the 70s and 80s but had largely abandoned in more recent decades, until it came back with a rush and is now the darling of every marketing department. Every project that involves any kind of machine learning (and many things that don't) will now be reported with 'AI' somewhere in the title of the article, even though it has nothing to do with ChatGPT, Claude, or Gemini.

And the reason is that, by appearing to have an understanding of natural language, generative LLMs have opened up the power of many of these systems to the non-technical general public, in the same way that the web browser in the 90s opened up the power of the Internet, which had also been in existence for decades beforehand, to ordinary users. (Many people ended up thinking the Web was the Internet, just as many people probably think ChatGPT has something to do with newspaper headlines about AIs diagnosing cancer.)

But it's not an analogy I'd like to push too far, because the technology of the World Wide Web did not invent new data, did not mislead people, did not presume to counsel them or tell them that it loved them. The similarity is that you needed to be something of an expert to make use of the Internet before the web, and you were therefore probably better able to judge what you might learn from it. If machine learning is statistics done badly, then 'AI' is machine learning made more unreliable, sounding much more plausible, and sold to the more gullible. Take any charlatan and give him skills in rhetoric, and you make him much more dangerous.

Regular readers will know that I am quite a cynic when it comes to most current uses of AI, and I consider myself fortunate that I was able to spot lots of its failings very early on. A few recent examples from ChatGPT, Gemini and other systems, some of which have been reported here, include:

  • Telling me that one eighth of 360 degrees was 11.25 degrees. (Don't trust it to do your financial planning!)
  • Telling a teenage friend that the distance from Cambridge to Oxford was 180 miles; she swallowed that whole and repeated it to me confidently. (It's actually more like 80 miles.)
  • Telling me that my blog was written by... well, several other people over the years, some of whom were flattering possibilities! (But there are several thousand pages here which all say "Quentin Stafford-Fraser's Blog" at the top.)
  • Suggesting a Greek ferry to a friend, as a good way to get to Santorini in time for our flight. (It didn't actually run on the days suggested, and we would have missed our flight if we had relied on it.)
And of course, the press has regular reports of more serious problems: So, some time ago, I announced Quentin's AI Maxim, which states that
"You should never ask an AI anything to which you don't already know the answer".
And for those who say, "But the AI systems have got a lot better recently!", I would agree. Some of my examples are from a few months ago, and a few months is a long time in AI. But I would also point out that, on Friday, when I asked the latest version of Claude to suggest some interesting places for a long weekend in our campervan, within about 2 hours' drive from Cambridge, one of its suggestions was Durham, which would probably take you twice that if you didn't stop on the way. I pointed this out, and it agreed.
"You're right to question that...I shouldn't have included it. Apologies for the error..."
Now, if I had been asking a human for suggestions, they might have said, "Mmm. What about Durham? How far is that from here?" But the biggest danger with these systems is that they announce facts just as confidently when they are wrong as when they are right, and they will do that whether you are asking about a cake recipe or about treatment for bowel cancer. Fortunately, I already knew the answer when it came to the suitability of Durham for a quick weekend jaunt! But here's the thing... Thirty-four years ago, I was very enthusiastic about two new technologies I had recently discovered. One was the Python programming language. The other was the World Wide Web. In both cases, more experienced research colleagues were dismissive. "It's not a proper compiled language." "We've seen several hypertext systems before, and none of them has really caught on." They were probably about the age that I am now. So, I don't want to be 'that guy' when it comes to AI. (Though I'm glad I *was* when it came to blockchains, cryptocurrencies and NFTs!) All of which brings to mind that wonderful quote from Douglas Adams:
"There's a set of rules that anything that was in the world when you were born is normal and natural. Anything invented between when you were 15 and 35 is new and revolutionary and exciting, and you'll probably get a career in it. Anything invented after you're 35 is against the natural order of things."
So in the last few weeks I have been doing some more extensive experiments with AI systems, mostly using the paid-for version of Claude, and the results have often been very impressive. They can be great brainstorming tools; I have to admit that some of the suggestions as to where we might go in our campervan were good ones... I'm just glad I didn't select the Durham option. They can be great search engines... just don't believe what they tell you without going to the source, or you too may have to call the coastguard. But perhaps I should modify the 2026 version of Quentin's AI Maxim to say something like:
You should never ask an AI anything where you don't have the ability, and the discipline, to check the answer.
And one of the areas where checking the answer can sometimes be an easier and more rigorous process is in the writing of software. I've been doing that a fair bit recently, and will write about that shortly. In the meantime, I leave you with this delightful YouTube short from Steve Mould. His long-form videos are always interesting - he has 3.5M followers for a good reason - and though I tend to avoid 'shorts' in general, this is worth a minute and half of your time.

CheatGPT and Alexander Pope

My friend Michael is a jolly good photographer, and I remember him telling me long ago, when he first started posting them online, that he'd had a comment from someone saying, "Your camera takes really nice photos!"

To which Michael had replied, "Thanks! Your keyboard writes really nice comments!"

Little did we know, back then, what was in store...

--

I have a sneaking suspicion that one of my clients is using a GPT to write some of his emails. I've had two or three in recent months that are just too carefully formatted: with nicely boldfaced section headings, too many bullet points, no typos, and they're just a bit too verbose: they read more like a legal document, a press release or a bad Powerpoint presentation than like a message to someone you've known for years.

My immediate reaction was that wonderful phrase I heard recently in an AI-related discussion: "Why would I want to read what somebody couldn't be bothered to write?" And if I knew that it was an LLM, and not a human, that had written it, that might have been my response. But I wasn't quite sure.

And this makes me think that accusing someone of using an AI, if in fact they haven't, could become a dreadful insult - I'd certainly take it that way. "You write like a machine." And, actually, one of the reasons I'm fairly confident that this particular chap is using it for some of his messages is that he also sends me missives which are much more human, and sound like him, and the difference is noticeable.

Unfortunately, kids aren't always smart enough to detect this distinction, and schools and colleges are finding they must now emphasise, to a much greater degree than in the past, that the essay a student produces for their assignment -- the end result -- has no value in itself. Your teacher isn't looking forward to receiving it because he really wants to have your great work of literature to keep on his bookshelf. No, it's the process of writing that essay that is the valuable thing, and doing so is the only thing that will help you when you're in the exam room at the end of the year without the help of ChatGPT (or 'CheatGPT' as I've recently heard it called). The recent idea that continuous assessment is a fairer way to assess students than the rather artificial world of exams is therefore being turned on its head.

--

In the early 18th century, Alexander Pope published his poem 'An Essay on Criticism', which introduced us to such phrases as 'A little learning is a dangerous thing', and 'Fools rush in where angels fear to tread'.

Think about that second one, for a moment . Can you imagine ChatGPT and its ilk ever coming up with that beautiful and succinct phrasing which incorporates so much human tradition and experience in just eight words? No, of course not. It might repeat it, if it had found it elsewhere, but it would never originate it.

AI systems are trained on the bulk of the data to be found on the internet, and they statistically predict what words and phrases might come next based on what they've seen most frequently. An AI's output will always be average, and never excellent. If you're lucky, then your AI will have been trained more on quality content than on the random output of the hoi polloi, and it might produce output which is in some senses slightly above average, but it is nevertheless always just plagiarising large numbers of humans.

Another famous line from An Essay on Criticism is the wonderful

To err is human; to forgive, divine.

And it was in the late 1960s -- yes, as early as that! -- that the newspaper columnist Bill Vaughan came up with a pleasing and much-quoted variation:

To err is human, to really foul things up requires a computer.

But I would suggest that we now need to revise that for our current age.

"To err is human, to excel is human, but to be truly average requires a computer."

Freedom of the press

Imagine an inkjet printer that you could repair yourself, and where you were actually encouraged to refill the cartridges!

That's the aim of the Open Printer project.

If the idea of printing from a roll of paper seems strange, you need to know that it incorporates a cutter too, so you can print single pages, and you're not restricted to one size of paper.

If, like many of us, you don't print very much these days, wouldn't it be nice to get the printer off your desk, and hang it on the wall?

I think this is a splendid idea, and I've signed up for updates (even though we already have three large printers in the house!)

I really hope they succeed.

Can Quentin get Quantum?

Like many... shall we say... classically-trained computer scientists (i.e. old ones), I have only the vaguest notion of how quantum computing actually works. My understanding of the various topics can be best pictured as a cloud-like set of probability distributions which doesn't exhibit any very high peaks!

So I was quite taken with Grant Sanderson's latest video in his '3Blue1Brown' YouTube channel, which does lovely graphical illustrations of mathematical concepts (each of which tends to get viewing figures measured in millions.) It increased my knowledge considerably of the kind of algorithms one might be able to run on a quantum machine.

"But what is quantum computing? (Grover's Algorithm)":

(Direct link to YouTube)

Yesterday

I liked this Beatles tribute, reposted on Mastodon but, it appears, written originally by one Sunni Freyer in the late 90s:

YESTERDAY

Yesterday,
All those backups seemed a waste of pay.
Now my database has gone away.
Oh I believe in yesterday.

Suddenly,
There's not half the files there used to be,
And there's a milestone hanging over me
The system crashed so suddenly.

I pushed something wrong
What it was I could not say.
Now all my data's gone
and I long for yesterday-ay-ay-ay.

Yesterday,
The need for back-ups seemed so far away.
I knew my data was all here to stay,
Now I believe in yesterday.

Misplaced trust

This might be a little technical for some readers, but don't worry, it's not actually the technical detail that's important...

On my home server, I run about half a dozen services that I need to access via a web browser, so they're all behind a Caddy reverse proxy which connects me to the right one, depending on the name I use in my browser: 'homeassistant', 'unifi', 'searxng', 'octoprint' etc. (All of these names are aliases for the same machine.)

One of these services is Nextcloud, which has user accounts, and I was thinking it would be handy if I could use those accounts to authorise access to the other services. Can I allow someone to use my web search frontend only if they have an account on my Nextcloud server, for example?

I thought I'd try out an AI system to see if it could speed up this process, because they're often good at this kind of thing - Google Gemini, in this case. And, to my delight, it gave me pages of detailed instructions.

It knew that Nextcloud supports the OpenID Connect system, told me how to set it up, and then how to use the oidc directive in the Caddy configuration file to connect the two, so that Caddy could ask Nextcloud whether the user should be allowed in. It gave me nice examples of oidc actually in use, and the parameters you'd need to configure when using it to talk to the Nextcloud instance.

"Great!", I thought, and grabbed a coffee, went upstairs to my machine, and started typing code to try it out. And it was then that I discovered...

Caddy doesn't actually have an oidc directive.

Five years before the iPhone

Trying to organise some of my old video footage recently, I came across a little demo I recorded of the AT&T Broadband Phone, a project we started in 1999 but which, sadly, died, along with the research lab that had created it, in 2002.

Looking back at it now, I notice how slow-paced it is compared to the typical YouTube video of today! So if you watch it, you might need a little patience! Nonetheless, it's quite fun to see some of the ideas we were considering back then, five or six years before the launch of the iPhone... things like the suggestion that streamed music "might be a service offered by a record company, where you pay a small amount for each track", for example...

Cordless Broadband phone and iPhone comparison

Direct link.

(P.S. I had an idea I had written about this here before... and indeed discovered that I had... but not since 2008, about eighteen months after the iPhone was launched.)

Coffee Pot - The Movie

For a long time, it has both bugged and bemused me that, though the first webcam ran for 10 years taking photos of our departmental coffee pot, there are almost no original images saved from the millions it served up to viewers around the world! I had one or two.

Then, suddenly, in a recent conversation, it occurred to me to check the Internet Archive's 'Wayback Machine', and, sure enough, in the second half of the coffeepot camera's life -- from 1996-2001 -- they had captured 28 of its images. I wrote a script to index and download these, and turned them into a slideshow, which you can find in my new and very exciting three-minute video: