this post was submitted on 25 Oct 2023
244 points (94.9% liked)

Programmer Humor

32453 readers
888 users here now

Post funny things about programming here! (Or just rant about your favourite programming language.)

Rules:

founded 5 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
[–] [email protected] 18 points 1 year ago (1 children)

Kenya give me a hint on what the joke is?

[–] [email protected] 13 points 1 year ago* (last edited 1 year ago) (1 children)

No joke here. Large language Models (which people keep calling AI) have no way of checking if what they’re saying is correct. They are essentially just fancy text completion machines that answer the question what word comes next over and over. The result looks like natural language but tends to have logical and factual problems. The screenshot shows an extreme example of this.

In general, never rely on any information an LLM gives you. It can’t look up external information that wasn’t in its training set. It can’t solve logic problems. It can’t even reliably count. It was made to give you a plausible answer, not a correct one. It’s not a librarian or a teacher, it’s an improv actor who will „yes, and“ everything. LLMs will often rather make up information than admit that they don’t know. As an easy demonstration, ask ChatGPT for a list of restaurants in your home town that offer both vegan and meat-based options. More often than not, it will happily make you a list with plausible names and descriptions but when you google them, none of the restaurants actually exist.

[–] [email protected] 1 points 1 year ago

Large language Models (which people keep calling AI)

As my AI teacher used to say, it's AI until someone builds it.