for legal responsibility hosting an instance, if someone posts illegal content, whatever, it should be possible to hide it from public access.
Asklemmy
A loosely moderated place to ask open-ended questions
Search asklemmy ๐
If your post meets the following criteria, it's welcome here!
- Open-ended question
- Not offensive: at this point, we do not have the bandwidth to moderate overtly political discussions. Assume best intent and be excellent to each other.
- Not regarding using or support for Lemmy: context, see the list of support communities and tools for finding communities below
- Not ad nauseam inducing: please make sure it is a question that would be new to most members
- An actual topic of discussion
Looking for support?
Looking for a community?
- Lemmyverse: community search
- sub.rehab: maps old subreddits to fediverse options, marks official as such
- !lemmy411@lemmy.ca: a community for finding communities
~Icon~ ~by~ ~@Double_A@discuss.tchncs.de~
For now you can do manual DB deletes for illegal content, but I'm working on purge recently to do the same thing through the API / UI. https://github.com/LemmyNet/lemmy/issues/904
I like the current system, where removed posts are blanked. I don't really see the point of preserving content after it's removed by a mod if the modlog already shows what got removed.
Given that purge exists, it seems like showing the content when you go to them makes sense.
The only other idea I have is implementing a click to reveal, but that seems like overkill to me.
Purge ( IE full database delete ) doesn't exist yet, but I've been working on it recently.
The danger with showing mod-removed content is that it still might not be a good thing to allow people to access it. IE lets say a post has a bunch of racist ranting, and a mod removes it. Currently that's being blanked out if you go to that post.
My mistake, I misunderstood and thought purge was already present.
I think my tendency would be to consider there to be about 3 categories of content. In parentheses is the result that I would probably go with when classifying the content in that category.
- Rule Violation (can be viewed upon click)
- Reprehensible (cannot be viewed)
- Banned/Illegal (purged from the DB)
Clearly banned/illegal or reprehensible content would also be rule violations in most cases, so I more mean that rule violation is the minimum category to be moderated and that a rule violation isn't excluded from the other categories, etc.
FBI or police of your country may require the material if illegal, but then, public access should be blocked ... i guess. Did you know Switzerland made apology of genocide illegal ?