this post was submitted on 18 Jun 2023
2 points (100.0% liked)

AI

139 readers
1 users here now

Artificial intelligence (AI) is intelligence demonstrated by machines, unlike the natural intelligence displayed by humans and animals, which involves consciousness and emotionality. The distinction between the former and the latter categories is often revealed by the acronym chosen.

founded 3 years ago
 

With the rapid advances we're currently seeing in generative AI, we're also seeing a lot of concern for large scale misinformation. Any individual with sufficient technical knowledge can now spam a forum with lots of organic looking voices and generate photos to back them up. Has anyone given some thought on how we can combat this? If so, how do you think the solution should/could look? How do you personally decide whether you're looking at a trustworthy source of information? Do you think your approach works, or are there still problems with it?

you are viewing a single comment's thread
view the rest of the comments
[โ€“] howrar@lemmy.ca 1 points 1 year ago* (last edited 1 year ago) (1 children)

the way that Bing has very clearly left footnotes for it's sources [...] is probably the best way forward at the moment

This brings us to the issue of being reliant on one entity (Bing) to decide whether the source is reliable. How do we know if this entity can be trusted, and how can we know if that ever changes? Assuming we can trust them, this just passes the problem onto someone else. How would this entity decide whether sources are reliable or not before feeding them to us?

Can you elaborate a bit on what you mean by metadata on generated images? What kind of metadata and what can you do with them?

[โ€“] fruitywelsh@lemmy.ml 1 points 1 year ago

Maybe take the same UX but apply it via wikidata instead?