At various times, most social media platforms have received criticism for alleged failure to prevent distribution of copyright-infringing content. Few, however, have been threatened with widespread blocking more often than Telegram. In a row that seemed ready to boil over last year, Telegram was given an ultimatum by the Malaysian government; come to the negotiating table or face the consequences. A Malaysian minister now says that Telegram is ready to fight piracy.
How much responsibility would a service like Signal have, if they were to inadvertently host a private group for pirated content? I believe signal groups can have up to 1000 members, and these members can be pretty anonymous given the need to only share an ephemeral username which can not be linked to a phone number or any other identity? Can they claim plausible deniability and not do anything?
IANAL and all the other anals, but my understanding is Signal wouldn’t be liable and wouldn’t have to do anything. They designed their service so they can’t know the content of the messages, so if a third party Maloyse (see what I’m doing there?) is reporting a message between Alice and Bob that Maloyse thinks to be illegal, Signal would be within legal grounds to bring into question how did M got that message, and it can’t be used as proof against Signal because there is no legal mechanism by which Signal could have acquired that message and act upon it - in fact, Signal has grounds to suspect Maloyse is crafting those messages, since neither Alice nor Bob have reported such message.
This post is licensed under CC BY-NC-SA 4.0. Feel free to contact me to negotiate for an alternative license.
This is the correct answer. Both Alice and Bob approve.
In a group chat, M wouldn’t be a 3rd party.
Maloyse absolutely can:
BobAlice’s homeYes, but we’re discussing group chats disseminating piracy links. Do you think it’s harder to join such a group chat and report it to signal than it is to do all the cloak and dagger nonsense?
Which one is harder has zero relevance upon how much work Signal has to do to vet the contents transmitted on the channels, which is zero, nil, because they can’t. Even if it was Charlie who reported the channel, Signal intentionally has no practical means to verify neither the accused contents nor the authenticity of the report. And this is actually good.
Infrastructure-wise, Signal (mostly) limits itself to only being a carrier. In a just world, a carrier who has been set up to take the limited responsibility of a carrier is not liable for the contents of carried things that are protected so that the carrier can not peek into. Sure, they can be legally pressed to change that and “upgrade” their lawyer plan to “content vetter”, but as far as I know that hasn’t happened yet.
I don’t get what you’re trying to say at all. If a party is in a group chat and reports it, they can provide their credentials to Signal to enable Signal to view the contents of the chat.
Yes, they’re a carrier that does not know the content of what they carry. But once they are made aware, the legal system considers them to now bear responsibility if they don’t take action. Whether or not that’s fair is a pretty large topic, though I’m inclined to think so myself.
What if that happens on Session? Can the nodes operators be sued even though they have no access to the content?