Very Artificial Intelligence

Regular readers might assume that I spurn all things AI-related, and this is not the case. I do use and occasionally derive benefit from the tools that tend to come under this all-embracing phrase du jour. But it does sometimes seem as if, for general questions, the LLMs can throw up wrong answers as often as right ones, and, like others, I then feel a compulsion to point this out. I had one week recently where three different ‘AI’ systems, in three completely different contexts, gave me three answers that were demonstrably wrong on key points.

So it’s important to know the answer before you ask them a question, in which case…

The real problem is the certainty with which assertions are made. There’s no expression of doubt, no humming and hawing, before the response.

Having been listening on my recent travels to the wonderful audiobooks of Patrick O’Brian’s stories, I thought it would be fun to ask ChatGPT about archaic naval terminology – did it know what was meant by ‘Three points off the larboard bow’.

It started with a cheery phrase along the lines of “Ah! Nautical terms! Always a source of interest.” and then went on to give various bits of authoritative information, even with helpful diagrams, but including the assertion that ‘A point in this context is one-eighth of a compass’s 360 degrees, so one point is 11.25 degrees.’

Mmm. Do a little calculation and you’ll see the problem.

I’m paraphrasing slightly, because when I went back to copy and paste the exact text, it had conveniently lost its history, and when I asked again, it made no mention of the ‘one-eighth’. (There are actually 32 ‘points’ on a compass, so the 11.25 degrees bit was correct. Four points are 45 degrees.) It’s a worrying thought: perhaps the system could detect that I was surprised and go back and cover its mistakes!

Anyway, it did get the rest of it right. ‘Larboard’ is an old word for ‘port’, which was abandoned in the mid-19th century because it sounded too similar to ‘starboard’. If yelled from the mast-top in the heat of battle in a gale, I guess this could cause problems. So ‘three points off the larboard bow’ means (roughly) 34 degrees to the left of the direction in which the ship is pointing.

Many, many years ago, when a couple of the projects I was working on started to get some press coverage, I remember noticing that, in general, every single article about my work contained errors. (There were a couple of exceptions — the Economist was one — but they were notable for their rarity.) These were generally unimportant mistakes, and they were made by well-intentioned journalists working for reputable papers back in the day when they could afford to do real research but who were, after all, humans. It was an important lesson for me in my youth: you can detect the errors in the things you know about, but remember that every article you read about anything is probably similarly flawed.

This is why I think it is important to keep taking note of the times we can detect wrong answers, because they will also happen in places that we cannot so easily detect. I wonder if adding more human-like phrases to the output of LLMs will enable us to take what they say with a pinch of salt, too? (Sea salt, of course.)

Enjoyed this post? Why not sign up to receive Status-Q in your inbox?

1 Comment

[…] One should always have something sensational to read on the net… « Very Artificial Intelligence […]

Got Something To Say:

Your email address will not be published. Required fields are marked *

To create code blocks or other preformatted text, indent by four spaces:

    This will be displayed in a monospaced font. The first four 
    spaces will be stripped off, but all other whitespace
    will be preserved.
    
    Markdown is turned off in code blocks:
     [This is not a link](http://example.com)

To create not a block, but an inline code span, use backticks:

Here is some inline `code`.

For more help see http://daringfireball.net/projects/markdown/syntax

*

© Copyright Quentin Stafford-Fraser