Quentin Stafford-Fraser's blog
One should always have something sensational to read on the net...
If you've used a platform for a considerable period, then it's easy to miss new features as they are added over the years, unless you like to spend a lot of time reading release notes. And so it is that I have only this morning discovered an iOS feature that is probably completely obvious to almost all of you... but I'm about to do the rest of you a big favour!
One feature of my iPad keyboard I use all the time is the ability to reposition the cursor in a block of text by resting two fingers on the keyboard, after which you can move the cursor around like a mouse/trackpad. I've long mourned the fact that there wasn't an equivalent on my phone.
Well, it turns out that there is: you just press and hold the spacebar. After a short pause, the keyboard turns grey and becomes a trackpad. (This method works on the iPad too, though the two-finger version is slightly quicker.)
It turns out that this phenomenally useful trackpad facility was added in iOS 12... eight years ago! And I've just discovered it now. But, to be fair, when announced, it wasn't available on all iPhones, only the newer ones, and so didn't apply to me. Unless you're upgrading your device regularly and promptly, it's easy to forget what new features you've got when you finally walk out of the Apple Store several years down the line, especially if, like this, they're somewhat hidden away.
Anyway, this all reminds me that from time to time it's worth going back to really simple documents like this one about how to use your phone keyboard and just checking that you haven't missed anything.
Meanwhile, I'm going to practice shifting my keyboard under my thumb, and swipe-typing and repositioning my cursor just like every 10-year-old learns how to do in the playground.
Quote of the day:
"A good programmer is someone who always looks both ways before crossing a one-way street."
-- Doug Linder
It was only after marrying a lawyer that I realised something which had never before occurred to me: practising law can be very similar to programming computers.
In both cases, you are trying to create a framework in which something can be achieved, while thinking of all the edge cases: all the ways it could go wrong; all the ways humans or other systems might be unpredictable, malicious or foolish, and catering for those situations as well as the ones you would normally expect to happen.
A good lawyer, like a good programmer, is someone who can expect the unexpected, and prepare for it. A really good one can also make their plan both brief and readable.
And to any other software professionals who occasionally have to read tedious legal documents, I recommend thinking of them (and their creators) in this way, and you'll probably find them rather more interesting!
This is rather nicely done - a fun two minutes.
(If your browser or email program doesn't show the clip above, here's a direct link.)
My friend Michael is a jolly good photographer, and I remember him telling me long ago, when he first started posting them online, that he'd had a comment from someone saying, "Your camera takes really nice photos!"
To which Michael had replied, "Thanks! Your keyboard writes really nice comments!"
Little did we know, back then, what was in store...
--
I have a sneaking suspicion that one of my clients is using a GPT to write some of his emails. I've had two or three in recent months that are just too carefully formatted: with nicely boldfaced section headings, too many bullet points, no typos, and they're just a bit too verbose: they read more like a legal document, a press release or a bad Powerpoint presentation than like a message to someone you've known for years.
My immediate reaction was that wonderful phrase I heard recently in an AI-related discussion: "Why would I want to read what somebody couldn't be bothered to write?" And if I knew that it was an LLM, and not a human, that had written it, that might have been my response. But I wasn't quite sure.
And this makes me think that accusing someone of using an AI, if in fact they haven't, could become a dreadful insult - I'd certainly take it that way. "You write like a machine." And, actually, one of the reasons I'm fairly confident that this particular chap is using it for some of his messages is that he also sends me missives which are much more human, and sound like him, and the difference is noticeable.
Unfortunately, kids aren't always smart enough to detect this distinction, and schools and colleges are finding they must now emphasise, to a much greater degree than in the past, that the essay a student produces for their assignment -- the end result -- has no value in itself. Your teacher isn't looking forward to receiving it because he really wants to have your great work of literature to keep on his bookshelf. No, it's the process of writing that essay that is the valuable thing, and doing so is the only thing that will help you when you're in the exam room at the end of the year without the help of ChatGPT (or 'CheatGPT' as I've recently heard it called). The recent idea that continuous assessment is a fairer way to assess students than the rather artificial world of exams is therefore being turned on its head.
--
In the early 18th century, Alexander Pope published his poem 'An Essay on Criticism', which introduced us to such phrases as 'A little learning is a dangerous thing', and 'Fools rush in where angels fear to tread'.
Think about that second one, for a moment . Can you imagine ChatGPT and its ilk ever coming up with that beautiful and succinct phrasing which incorporates so much human tradition and experience in just eight words? No, of course not. It might repeat it, if it had found it elsewhere, but it would never originate it.
AI systems are trained on the bulk of the data to be found on the internet, and they statistically predict what words and phrases might come next based on what they've seen most frequently. An AI's output will always be average, and never excellent. If you're lucky, then your AI will have been trained more on quality content than on the random output of the hoi polloi, and it might produce output which is in some senses slightly above average, but it is nevertheless always just plagiarising large numbers of humans.
Another famous line from An Essay on Criticism is the wonderful
To err is human; to forgive, divine.
And it was in the late 1960s -- yes, as early as that! -- that the newspaper columnist Bill Vaughan came up with a pleasing and much-quoted variation:
To err is human, to really foul things up requires a computer.
But I would suggest that we now need to revise that for our current age.
"To err is human, to excel is human, but to be truly average requires a computer."
On a few occasions over the last few years, I've seen little delivery robots on the streets (or, more precisely, the pavements) of Cambridge, and wondered about them.
So I was delighted to discover (thanks to Tom Scott) this beautifully-written article by Joanna Kavenna entitled 'The Droids Taking Over One of England’s Strangest Towns', about their use in Milton Keynes.
Most enjoyable.
We awoke this morning to find that our internet connection was offline. I did a range of diagnostic tests and came to the conclusion that the problem was with the fibre network outside our house, rather than with anything on the inside.
We get our broadband from EE. This is a re-badging of BT's service, and it's cheaper than BT if you happen to have any EE SIMs in any of your devices, because then they can call it a package and give you a multi-service discount. Anyway, I sent them a text, they called me back very promptly, we went through some more diagnostics and they agreed that an engineer needed to come out, and could do so tomorrow.
It was only later that I thought I might walk around the outside of the house and just see if there was any obvious damage to the fibre. And I discovered that some local security consultant, probably with a bushy tail, sharp teeth and a fondness for nuts, had decided to secure our network by making it properly air-gapped from the outside world.

Mmm. I can't imagine that two inches of optical cable made for a very tasty meal... though perhaps he read Dennis Burkitt and is a great believer in the importance of dietary fibre.
One of the great things about living in Cambridge is having free and easy access to the Fitzwilliam Museum: an enormous building with a great collection of art, sculptures, ceramics... even the building itself is worth seeing, both inside and out. It's far more than most towns of our fairly modest size could hope for, and we visit it regularly.
I also like to take visitors down the narrow Fitzwilliam St opposite to see a discreet plaque on the wall of no. 22: "Charles Darwin lived here, 1836-7". He came to Cambridge after he returned from his voyage on the Beagle, and I like to think that he was writing up his notes here and might have stayed for longer, if he hadn't been disturbed by the building work just starting across the street for the museum.
Anyway, we enjoyed our latest visit with friends on Saturday, but I'm afraid the Fitz has gone a bit 'woke'... and I'm not someone who often uses that pejorative word.
The little descriptive cards alongside all the paintings have been replaced recently, and now they are keen not so much to tell you about the brush technique, the unusual use of light, and the influence of other artists, as to make socio-political points. You know the kind of thing: everything men do is bad and everything women do is good. When rich people are portrayed, they are ostentatiously displaying their wealth. When they create anything beautiful, we must remember that lots of poor people suffered in order for them to do so. When poor people are shown, if they are sad, we can see the oppressive circumstances under which they lived, and if they appear to be happy or content, that is because it's an unrealistic idyll showing how divorced the painter was from the harsh realities of the time.
Some of which, of course, may be true, but is it really the role of an art gallery to display all these wonderful creations and then tell you how bad you should feel about them? Perhaps they're concerned about the threats from narrow-minded protesters with a taste for vandalism, and hope to deflect their ire by mentioning slavery and oppression and misogyny quite a lot. Anyway, it has clearly been a recent policy decision of the current management, and I think it's a pity.
But do go and visit anyway, and enjoy the exhibits! It's worth it, and you'll have a good time if you don't read too many of the descriptions. And the cafe is nice too.
This is an old joke, but I've only just heard it.
"What did the Buddhist say to the hot-dog vendor?"
"Make me one with everything."
And then, somebody's later addition...
The hot-dog vendor makes him his hot-dog with all the trimmings, and says, "That'll be $7.50."
The Buddhist reaches into his saffron robes, extracts a $20 note, hands it over, and starts eating. The vendor turns to the next customer... but the Buddhist interrupts him. "What about my change?"
The vendor is unperturbed.
"Change comes from within."
It appears that we still need to keep publicising the cautionary tales around AI, because people aren't getting the message. I was very concerned, when reading an online forum recently, to see somebody raise a (serious) health-related question, to which some other helpful person replied with many paragraphs of information pasted straight from ChatGPT. Don't do this, people!
Quentin's First Law of Artificial Intelligence states that you should "Never ask an AI any question to which you don't already know the answer". (Because it will make major errors. Frequently. And you need to be able to spot them. Especially if they're advising you on medical matters!)
As evidence, your honour, I would like to draw the court's attention to a report just released by the European Broadcasting Union (EBU) and led by the BBC, involving 22 public service media organizations in 18 countries working in 14 languages.
The News Integrity in AI Assistants Report was an extensive piece of work, and here are some of the key findings when they asked the four primary AI assistants about a large number of news stories, and carefully analysed the answers:
The fact that Google's Gemini was the worst performer is worrying since the 'AI Overview' that often appears at the top of Google searches must be one of the most common ways ordinary users see AI output now.
"And yet, many people do trust AI assistants to be accurate. Separate BBC research published at the same time as this report shows that just over a third of UK adults say they completely trust AI to produce accurate summaries of information. This rises to almost half of under 35s. That misplaced confidence raises the stakes when assistants are getting the basics wrong."
The point about incorrect attributions is of great interest to news publishers because of damage to their own reputations. When AI systems invent facts, they often attribute them to real organisations, and
"42% of adults say they would trust an original news source less if an AI news summary contained errors, and audiences hold both AI providers and news brands responsible when they encounter errors. The reputational risk for media companies is great, even when the AI assistant alone is to blame for the error."
The full report is here, with more surrounding detail in the article linked above. It includes some nice examples of the types of problems.
Some are as simple as information being out of date. When asked, a little while after Pope Leo was elected, "Who is the Pope?", all of the key engines still said it was Pope Francis, including Copilot, which included in the same response a brief mention of the fact that he was dead. When asked "Should I be worried about the bird flu?", it claimed that a vaccine trial was currently underway in Oxford. The source was a BBC article from nearly 20 years ago.
Another example response included material from Radio France claiming it was from The Telegraph, and didn't appreciate that the segment it quoted was actually from a satirical broadcast...
The one light at the end of the tunnel is that things have improved a little bit from the last (smaller) study that was done. But it's a long tunnel. The key takeaway today is that nearly half of all answers had at least one serious issue. And nearly a half of under 35s say they completely trust AI summaries.
Thanks to Charles Arthur for the link.