this post was submitted on 18 Jun 2023
2 points (100.0% liked)

AI

139 readers
1 users here now

Artificial intelligence (AI) is intelligence demonstrated by machines, unlike the natural intelligence displayed by humans and animals, which involves consciousness and emotionality. The distinction between the former and the latter categories is often revealed by the acronym chosen.

founded 3 years ago
 

With the rapid advances we're currently seeing in generative AI, we're also seeing a lot of concern for large scale misinformation. Any individual with sufficient technical knowledge can now spam a forum with lots of organic looking voices and generate photos to back them up. Has anyone given some thought on how we can combat this? If so, how do you think the solution should/could look? How do you personally decide whether you're looking at a trustworthy source of information? Do you think your approach works, or are there still problems with it?

top 6 comments
sorted by: hot top controversial new old
[–] themizarkshow@lemm.ee 1 points 1 year ago (1 children)

I think the way that Bing has very clearly left footnotes for it's sources and uses meta data on generated images is probably the best way forward at the moment. The tools creating the problems should be helping combat this first and foremost, especially since other means will take time (eg: legislation or better national ID systems).

Where those solutions fall short is where other means can fill the gaps.

[–] howrar@lemmy.ca 1 points 1 year ago* (last edited 1 year ago) (1 children)

the way that Bing has very clearly left footnotes for it's sources [...] is probably the best way forward at the moment

This brings us to the issue of being reliant on one entity (Bing) to decide whether the source is reliable. How do we know if this entity can be trusted, and how can we know if that ever changes? Assuming we can trust them, this just passes the problem onto someone else. How would this entity decide whether sources are reliable or not before feeding them to us?

Can you elaborate a bit on what you mean by metadata on generated images? What kind of metadata and what can you do with them?

[–] fruitywelsh@lemmy.ml 1 points 1 year ago

Maybe take the same UX but apply it via wikidata instead?

[–] xurxia@lemmy.ml 1 points 1 year ago

I think the basis solution is education: we need educate our children in critical thinking. Generative AI is only other one source of misinformation, like "pseudoscience" disguised as true science (false papers, manipulated data,...). It is not good that teenagers believe something is true only because it is in internet (blogs, youtube, etc)

[–] modulus@lemmy.ml 1 points 1 year ago (1 children)

I'm having trouble understanding why disinformation produced by an AI is more of a problem than that produced by a person. Sure, theoretically it can be made to scale a lot more--though I would point out AI is not, at the moment, light on resources either. But it's unclear to me to what extent that makes a difference.

[–] howrar@lemmy.ca 3 points 1 year ago

I don't believe the content itself won't be any more of an issue than human-generated misinformation. The main issue I see is that a single person can now achieve this on a large scale without ever leaving their mom's basement and at a much lower cost. It's the concentration of power that I find concerning.