this post was submitted on 29 Jun 2023
31 points (100.0% liked)

Technology

37738 readers
48 users here now

A nice place to discuss rumors, happenings, innovations, and challenges in the technology sphere. We also welcome discussions on the intersections of technology and society. If it’s technological news or discussion of technology, it probably belongs here.

Remember the overriding ethos on Beehaw: Be(e) Nice. Each user you encounter here is a person, and should be treated with kindness (even if they’re wrong, or use a Linux distro you don’t like). Personal attacks will not be tolerated.

Subcommunities on Beehaw:


This community's icon was made by Aaron Schneider, under the CC-BY-NC-SA 4.0 license.

founded 2 years ago
MODERATORS
 

There is huge excitement about ChatGPT and other large generative language models that produce fluent and human-like texts in English and other human languages. But these models have one big drawback, which is that their texts can be factually incorrect (hallucination) and also leave out key information (omission).

In our chapter for The Oxford Handbook of Lying, we look at hallucinations, omissions, and other aspects of “lying” in computer-generated texts. We conclude that these problems are probably inevitable.

top 17 comments
sorted by: hot top controversial new old
[–] snake_case@feddit.uk 13 points 1 year ago (5 children)

If you're looking for a factual response from chat GPT you're using it wrong. It's designed to produce text that looks correct. It's not a replacement for Google or indeed proper research. For more on this watch the leagle eagle video on chat gpt case: https://youtu.be/oqSYljRYDEM

[–] trachemys@iusearchlinux.fyi 4 points 1 year ago

But it sure knows how to sound like it is authoritative. Very convincing.

[–] Dr_Cog@mander.xyz 2 points 1 year ago

It's decent for parsing text, given you are careful about the prompt generation.

I am exploring it's use in assessing speech-based cognitive assessments, and so far it is pretty accurate

[–] bionicjoey@lemmy.ca 2 points 1 year ago

Unfortunately, fucking everyone is treating it like it knows things.

[–] Karlos_Cantana@sopuli.xyz 1 points 1 year ago

I was trying to use it to fix coding, but I could never get it to work right.

[–] Spudger@lemmy.sdf.org 1 points 1 year ago

I'm not using ChatGPT at all.

[–] furrowsofar 3 points 1 year ago (2 children)

Oh please, let's stop humanizing this stuff by calling this stuff hallucinations, and lying. These hidden boxes are just big table looks that use some sort of fancy interpolation/extrapolation engine. They are not human or intelligent yet if ever. It seems to be that the people pushing and talking about this stuff that are ones "hallucinating".

[–] zzzzz 2 points 1 year ago (2 children)

These hidden boxes are just big table looks that use some sort of fancy interpolation/extrapolation engine

Could the same not describe the human brain?

[–] Lowbird 2 points 1 year ago* (last edited 1 year ago) (1 children)

It knows the probabilities that one word or sentence will follow another, based on the data it was trained on. That's it. It's like typing exclusively through a smartphone keyboard's predictive text, just much more complex and trained on a much larger dataset.

It has no way of connecting any word or sentence to what that word or sentence means or represents in the real world, and it's absolute pants at making up anything new, which is why AI generated stories are cliches from top to bottom.

Like, it knows what words and phrases are associated with the word cat, or with a description prompt like "a cute story about your cat", so it can imitate that text, or imitate it and mix it with imitations of other text patterns (" give me a cute story about your cat riding a dinosaur in a sarcastic tone"), but it's just a mirror that reflects back what humans put into it.

If it must be compared to a human brain, it's more like exclusively broca's area of the human brain - the part that handled language. If broca's area is damaged in a head injury, a human can end up with aphasia, unable to speak coherently but maybe still able to understand others perfectly well , or able to speak but unable to understand language from others, or both, but in all other ways the person can be completely fine and still able to live and solve puzzles and interact with other humans normally. (Granted, I am not a neurologist, and I'd imagine head injuries like these can probably get much more complicated.)

Anyway, hypothetically, if you were to take broca's area out of a human brain and keep it alive in a jar on its own, without the rest of the brain: I don't think it could be described as sentient or thinking on its own. Although the human brain is so complex and interconnected and poorly understood (beyond what area is generally associated with what function) and malleable/plastic based on experience (e.g. the brains of blind people re-purpose areas normally used for sight to he used for other purposes instead, absent visual input) that I could be wrong / it could be arguable.

The experiences of people who have had the two halves of their brain separated (this used to be a treatment for seizures, apparently), who then seem to have the two halves thinking and operating independently, or of people who manage to live largely normal lives despite missing half or more of their brain entirely, would suggest that maybe even a tiny piece of a brain could be a thinking being, I suppose.

There was an experiment a while back... MIT, maybe? Like a decade or more ago. Where they made little robots that had a kind of scaffold/medium in which they planted and grew neurons sourced from mouse brains. The robot brains, iirc, changed in reaction to the input the robots were given/their environment and sensors, and they showed personality differences in the way they behaved.

LLM's don't do anything like that, though, so far as I know. They don't restructure themselves in response to input like an organic brain can. They're a really complicated pile of if-thens, and while the inputs and therefore outputs change, the mechanism that turns inputs into outputs never does.

This comment is like 90% tangent at this point. Oh well.

Anyway, there's a really good classic metaphor for this that explains it much better, about a man in a locked room who receives mail in a language he can't speak through a slot in the wall, but he is still able to sort the mail in complex ways even though he can't understand the content of any of it. I can't remember the whole thing or the name of it off the top of my head though. I'll try to remember.

[–] zzzzz 1 points 1 year ago

Thank you so much for your thoughtful response. I'm sorry for not seeing it for so long! If you can believe it, I just discovered the "inbox" in my lemmy app and am going through all the things people said to me over the past month.

This whole topic is really interesting to me. I hear what you're saying and imagine the distinctions you're drawing between these models and real brains are significant. I can't help but wonder, though, if we, as humans, might be poorly equipped to recognize the characteristics of emerging intelligence in the systems we create.

I am reminded vaguely of the Michael Crichton book Andromeda Strain (it has been many years since I read it, granted) wherein an alien lifeforms based on silicon, rather than carbon, was the major plot object. It is interesting to think that something like an alien intelligence might emerge in our own networked systems without our noticing. We are waiting for our programs to wake up and pass the Turing test. Perhaps, when they wake up, no one will even see because we are measuring the wrong set of things...

[–] furrowsofar 1 points 1 year ago

My issue is using humanizing language which adds a lot of baggage and implied assumptions which are miss leading. Bottom line is that these things just give crazy results at times through design and the data you put it and even at best they are rather dumb. Maybe useful but the user has to know their limitations and verify their results.

[–] bermuda 1 points 1 year ago (1 children)

Oh please, let’s stop humanizing this stuff by calling this stuff hallucinations, and lying

do you have a better word for it?

[–] furrowsofar 2 points 1 year ago (1 children)

I would call it generating unexpected useless and perhaps dangerous and misleading results. There is no motive like lying. There is no medical condition like hallucinations.

[–] Lowbird 2 points 1 year ago (1 children)

People already use "lying" when talking about other confusing inanimate objects/interfaces that don't have motivations (ignoring the motivations of their creators that may seep through). Examples:

"Google maps lied to me and I thought I was two blocks west of where I was."

"The fuel gauge says the tank is empty but that's a lie."

"These chip bags look full but it's all lies, they're just full of air."

It's harder to think of examples for "hallucinate", though people do describe things as " halluconatory" or "an acid trip" and so on.

I think even in world where everyone understood that LLM's do not think or understand, and where everyone understood how they generate their results, I think people would still talk like this.

I understand the frustration, but also it seems as tall an order to me as asking people not to personify their desktop computers and phones and other inanimate objects, or to not apply pronouns other than 'it' to stuffed animals and dolls. This kind of personification and personifying metaphoric language is one of those things humans are just inclined to so, however inconveniently in this case, imo.

[–] furrowsofar 1 points 1 year ago

I am really more worried about the researchers and the product people getting ahead of themselves on one hand and on the other people not understanding the huge limitations of these things at the moment. Essentially people not being skeptical when they use the tech.

[–] MagicShel@programming.dev 1 points 1 year ago (1 children)

If your use case can't handle hallucinations, NLP is not a good fit.

[–] Spudger@lemmy.sdf.org 3 points 1 year ago

Most people don't have "use cases". They just want lazy solution to a problem that requires some actual thought. Never forget that half the population is below average intelligence. Also, there's a reason tabloid newspapers still sell the most copies: people.