Microsoft says bizarre travel article was not created by ‘unsupervised AI’

According to a recent article posted by Microsoft Travel on, attractions worth checking out on a visit to the Canadian capital of Ottawa include the National War Memorial, Parliament Hill, Fairmont Château Laurier, Ottawa Food Bank … hang on, Ottawa Food Bank?

Spotted in recent days by Canada-based tech writer Paris Marx, the article puts Ottawa Food Bank at number 3 in a list of 15 must-see places in the city. And as if that wasn’t bad enough, the accompanying description even suggests visiting it “on an empty stomach.”

The piece was originally thought to have been created by generative artificial intelligence (AI). But Microsoft has since said that “unsupervised AI” was not involved. The entire article has now been taken down, though you can view an archived version of it.

Here’s the food bank description in full:

“Ottawa Food Bank — The organization has been collecting, purchasing, producing, and delivering food to needy people and families in the Ottawa area since 1984. We observe how hunger impacts men, women, and children on a daily basis, and how it may be a barrier to achievement. People who come to us have jobs and families to support, as well as expenses to pay. Life is already difficult enough. Consider going into it on an empty stomach.”

A screenshot of a Microsoft travel article.

With its insertion into the travel article clearly an error — and an awful one at that — it seemed probable  that the piece was knocked together using generative AI, a technology that we know Microsoft has a huge interest in.

But a statement from the company claimed the issue was “due to human error,” adding that it was “not published by an unsupervised AI” but instead generated through “a combination of algorithmic techniques with human review, not a large language model or AI system.”

Still, the company appears to have fallen short on several counts. First, it failed to perform proper human checks on the article before posting it, and second, nowhere on the webpage did it say that the content was generated through “algorithmic techniques.”

The mishap demonstrates the continuing need for human oversight when technology is used to create content. Slip-ups can be costly, as evidenced by a recent case in New York City in which a lawyer used the AI-powered ChatGPT chatbot to find examples of legal cases that he then included in a document to support a client’s case. But it was later found that ChatGPT had made them all up.

This article has been updated to include details of Microsoft’s statement.

Editors’ Recommendations

  • Meta is building a space-age ‘universal language translator’

  • Google Bard could soon become your new AI life coach

  • Hackers are using AI to create vicious malware, says FBI

  • Top authors demand payment from AI firms for using their work

  • GPT-4: how to use the AI chatbot that puts ChatGPT to shame