But all coming on here and saying "ooohh, this is bad, innit!" is not very interesting, and unlikely to prevent it.
Why is the EU doing this? Which political groups are supporting and opposing it? Why now? How are vendors responding? How does it affect non-EU countries?
https://balkaninsight.com/2023/09/25/who-benefits-inside-the...
"Oui oui!"
I disagree - this is how the internet can strengthen democracy.
Upvoting and commenting makes this post hit the top of HN and stay there. This makes it visible to many EU citizens who can reach out to their MEP's to ask them to vote against it. Seems a pretty effective strategy to me as someone living in a non-EU country.
Although agree that we should also be discussing the questions you raised.
I'll state that as an American I'm quite unhappy with this as I know the regulations will also affect me and the truth of the matter is that I have a much smaller voice in this matter due to not being a European citizen. I do have additional worry since it was not that long ago which we saw the results of authoritarianism in Europe (though it did result in the strengthening of my country). And my concern is that authoritarianism creeps, often with good intentions but poor foresight. My biggest fear is that we did not learn the great lesson from WW2, in that Germany did not in fact go from good people to the entire country being evil and back to being good people. If we can't understand this process and see how it actually happens (with the details) it will only repeat, led by people that have. But I don't know how to get people to understand subtleties, and that seems like a major issue in a world growing increasingly complex.
Edit: here it is
https://netzpolitik.org/2022/dude-wheres-my-privacy-how-a-ho...
Before CA, the received wisdom was that if you do something bad, you will need to resign before you are pushed for causing damage to the organisation reputation and therefore electability. This was perhaps borne out with enormous error bars by focus groups and polls asking "would you still vote for X in case of Y".
After CA, and in particular the live social media sentiment data that was gathered around the debacle of the UK Brexit referendum, the data showed that actually egregious misbehaviour did not materially affect sentiments, and perhaps even appealed to a larger proportion of people than believed. For example, the famous "shy Tory" might not show up well in a focus group, but it all hangs out after analysing Facebook's data.
With that data in hand, people started doing things that they would never have dared to do before, knowing that it won't actually harm them, at least in the short run (since this data only shows short term effects).
And that's how we go from resigning over fairly small gaffes to the "screw it, what you gonna do, we know you won't vote for the others, we've seen your data" of today.
Not long ago, calling a woman a bigot on a hot mic was a dreadful PR disaster. Now, you can physically snatch a journalist's phone and it barely registers.
It does, however stack up over time with catastrophic final effects, much like chasing only quarterly figures or always postponing dealing with technical or real debt.
Before that change, a scandal in the papers also meant you had to have lost political favour with the people who owned the media companies, ie, were losing big political battles. You also had no hope of being re-elected through a hostile media because if they didn't carry a favourable message there was no way to communicate with voters. I'd argue people like Jeffery Epstein never really made it to trial or public attention because stories got buried.
Afterwards the better approach is to point and shout "Fake News". There are multiple channels that reach voters and it turns out that the corporate media are actually much more unreliable and unpopular than were previously suspected. A lot more dirty laundry is aired and the Streisand effect takes hold.
CA wasn't the change, it was just one of the first big scandals to happen in the new era.
* a 71 year old lady, with no social media and no public speeches ever.
* a guy who used a nickname for his last name, that matched with a military general (who is well known), and many people thought he was the general
* a "journalist" that was caught twice talking on live TV, conversing with a pre-recorded video
* a convicted criminal
It's impressive to manage to fail as a politician.
YouTube prankster voted in as Cyprus MEP - https://www.bbc.com/news/articles/c4nnrwr72dqo
Then why did so many vote extremist anti EU? There just has been EU election in germany and the nationalist gained a lot. And none of the big parties otherwise said a clear no to that, so what could I do, except vote a small party against that, but too small to really do something?
Voters care. But they see often no point in voting anymore.
But then there are the „Wahlhelfer“ who openly proud themselves in Twitter to invalidate votes for parties they disagree with.
This is democracy?
- https://id-party.eu/program/ (ID Party Official Site)
- https://idgroup.eu/news/online-censorship-is-a-threat-to-eur... (ID Group News)
- https://en.wikipedia.org/wiki/Identity_and_Democracy (Wikipedia Overview)
- https://id-party.eu/declaration-of-antwerp/ (ID Party Official Site)
The ID group is opposed to EU-wide surveillance measures, and promises to protecting individual privacy and national sovereignty.
Won't you take her back and make her Minister for Castles or something,
Is it beacuse a lot of people feel unsafe or is it because the people supposed to ensure our security see it as the "easiest" or most effective way to do their job?
Is there so much benefit to having a fuctioning mass surveilance apparatus, and if yes, who benefits of it if not the people for whom these rules for in the first place?
Democracy is a spectrum and it comes in many flavors.
For example nobody voted for von der Leyen, the French voted against the EU referendum in 2005 but the government still went with it, there hasn't been any referendum in France since
Democracy, even a flawed democracy leaves the status quo power structure vulnerable to being changed by popular political action. Mass surveillance allows the existing players to identify any nascent political movements that may eventually grow to threaten them and undermine or destroy these movements before they ever become a threat.
The issue is that Europe is behind in tech; particularly big communication networks (aka: social networks). One key element here is that the amount of taxes paid by the Facebooks and co. is ridiculously low and their importance in the economy is getting bigger and bigger. This causes a significant risk for the future.
Any attempt that we made to combat this on the technological aspect has been a complete failure.
To protect its citizen, Europe uses the biggest weapon at its disposal: regulations. The point is not to impose mass surveillance, nor is it to protect the children; the point is to hurt social networks because they are perceived as a threat (real or not). Hammer them with regulations until it's almost impossible to comply, if possible by implementing conflicting ideas (protect privacy of everyone BUT check every image for child pornography !).
The desired outcome is that: either the social network goes out of Europe, or decides to accept the fines, which more or less corresponds to what Europe believes should have been paid by a fair tax system.
Expect the exact same thing to happen with AI.
1. Voluntary chatcontrol (i.e. temporary derogation of the ePrivacy directive)
2. Mandatory chatcontrol (i.e. services must scan private communication once a detection order is issued)
The first version of chatcontrol is currently in effect, but it will expire in a few years. It was introduced, because social media (like facebook) was already scanning private communication to find CSAM and then someone pointed out that this is illegal in the EU and thus the ePrivacy directive was sabotaged to allow the scanning of private communication.
Facebook wanted to have a legal basis for the scanning of private communication. This does not hurt non-EU social networks, it helps them.
If the EU wished more competition by domestic companies, they could simply pass a law restricting the income tax by member states to not go over 50% of earned income. In my current jurisdiction, a lowly IT worker can easily pay a 66% tax rate, with 54% being paid as nominal taxes and the rest going as employer taxes. Unless you are self-employed, in which case you will get to pay your 66% with no smoke curtains in the middle.
Besides, there is a laaaarge amount of influential, rich and well connected people like Ashton Kutcher/Thorn willing to profit off of it. They're selling out our freedoms for personal gain.
Tech doesn't create terrorism or child abuse.
Technology is evil. Specifically, the Internet is evil.
It's why it's got such a uniquely rich potential for dystopia, why all news is always bad news, why it's always getting worse.
Technology, as it progresses, is the enabling of power. So you'll find more and more power exerted over you because people can't resist. You'll have to assert power of your own to counter it, like switching to Linux or using a VPN. Or torrenting a media to escape the oppressive DRM placed upon it. But it's essentially a war you've found yourself enlisted in. If you've got the chops to fight in it, you're lucky; most don't.
I single out the Internet because it connects us with oppressors and makes us reliant on them more intimately than ever before. AI will be able to extend that oppression even in an air gapped environment because now the oppressor's intent can be packaged up and installed on the machine like never before -- no connection required.
And, in the limit, I think it all ends with gray goo, Daybreak style.
I am not saying this is false, but think about the inverse: are there non-democratic states never drifting into mass surveillance? Maybe it is a symptom of a developed, high-trust society.
Which doesn't make mass surveillance a good thing. I'm just contemplating whether it's even possible to turn out different.
This is the point that needs to be hammered home. Allowing governments access to everyone's confidential information is a massive security disaster waiting to happen because bad actor's will target this backdoor.
Data privacy used to be trampled on with the fear of ‘terrorism’, but the Americans blasted the airwaves with the word so much that it diluted the word to the point it instills zero fear.
Now the new words of the times to trample on privacy is CSAM.
And like utter fools, the public fall for the same crap time and time and time again.
The truth is, the world will always be shit and have shit people in it. Those shit people will do shit things.
It’s a fight over your soul now. And the AI is going to love love love everything it has on every single person on the planet. I can’t wait for the AI to come for us all - we are collectively just awful (and I believe we the west are probably more awful to humanity than even the Chinese, Israel or the Russians, and that is saying something).
I am mad.
With a mechanism like this, this probably will increase further, apart of course for the unjust violation of privacy and even dignity. This is a law that contradicts the constitution very directly.
Also, this is policy that was not brought through democracy. It was created by European Commissioners that only have a very low democratic legitimacy for far reaching policies like this.
It would be a disaster for the EU and all its citizens if this comes through and everyone will loose.
Hint: the former - hundreds of thousands, the latter - zero
Literally soul destroying. As in life destroying.
Meta data kills people. Now it’s moving way further than meta data..
not other countless nations that have been enslaved, colonized, invaded, subjugated, stripped in the name of democracy and religion.
Russia, CCP etc are horrible too - but let us not forget history.
The Global South Remembers.
This is a wild POV to hold and one I'm pretty disappointed to hear on HN. You'd really prefer a world of Russian hegemony over NATO? My cousins in Ukraine would be shocked to hear otherwise smart people cheering on a regime which the rest of the civilized world has broadly condemned.
You are making bad assumptions here.
But if you want to put it into perspective… think about what is happening in Gaza. Now compare to Ukraine. Russia has plenty of missiles and by comparison (even while absolutely awful) have shown much more restraint. I don’t support them at all though!
I definitely agree with the article here. Probably after Chat Control will be implemented for CSAM this would act as a gateway towards using this tehnology for other things.
I am curious whether chat control will extend to mail or other means of online comunication, if it will be implemented ofc(hopefully not).
I’ll probably start communicating with my friends over phone more either way, I don’t want my conversations to be monitored 24/7.
Talking to people is great. I write science fiction for a hobby. In my stories, when two people want to have a private conversation to discuss some economic barter that can be construed as tax evasion, they take off all their clothing and go swimming to a beach with noisy waves. But there is always that lingering fear about if anybody surreptitiously got a microphone implanted during their latest root canal treatment...
Truth be told, I'm a very unimaginative bloke, because in my current jurisdiction banks are already forced to report on their customers, and in my previous one, the government had a decided phobia of cell-phones and attributed to computer printers in the hands of civilians the same dangers of an independent printing press.
Ironically the pervs will still be using WhatsApp, and just put their CSAM in a password-protected zip file before sending.
Even if you use an open-source clone without scanner, your contacts most likely will use an app with builtin scanner. Your communications will be scanned on their end.
At that point I'm wondering why we don't also open and scan regular mail at the post office before delivery.
However most users will be using the matrix.org homeserver, which makes it effectively centralized. Though I can still create my own homeserver that talks to matrix.org.
Would matrix.org be forced to offer scanning / a backdoor on the homeserver? Or would they be forced to add something to the official apps, which is pretty ineffective as there are many client apps.
All in all this proposal seems like a complete mess.
Amazing.
[0] https://op.europa.eu/en/web/who-is-who/organization/-/organi...
It's too easy to frame the issues in a context where I am right and the other part is wrong, because yes.
I used to run a service free as in speech and as in free beer to host little snippets of text to pass them around, I started it because I was tired of not being able to easily share some piece of information from my mobile phone to some other device.
It's astonishing the amount of spam and illegal material, from simple copyright infringement to porn - probably revenge porn or people sharing media of unaware victims-, that the service was getting used for in a very short amount of time.
I guess there are probably botnets scanning the entire internet for vulns or open services to abuse them.
I shut it down voluntarily, nobody reported me or said anything to me or forced me to do a thing, I simply did not want to be part of something like that and and had no time/resource to properly fight the spam.
This is the sad state of affairs and refusing to even discuss about it, to me means burying your head in the sand.
https://netzpolitik.org/2022/dude-wheres-my-privacy-how-a-ho... https://news.ycombinator.com/item?id=17914935
https://news.ycombinator.com/item?id=40713065
But it seems (as always) not easy to determine.
If we don’t hijack privacy in messaging, how do we fight crime happening on a message platform? If government doesn’t have access to message contents, what’s stopping criminals from using the platform and never get tracked down? Or proven guilty, since all the proof is safely encrypted? Aren’t we hurting ourselves by being so obsessed with privacy? Again, I apologize for ignorance and am curious
Think about it - if you're a criminal, and you know about chat control, why would you risk your chats being leaked at all? Why wouldn't you use a different app that you know to be more secure (this already happens for any serious crime already btw)
It's precisely the law-abiding people whose privacy will be invaded for no-gain
That's the only way to enforce it and there is history for it in cryptography export laws and anti-circumvention laws.
Compare it to: if we don't put cameras and microphones in everybody's houses, how do we fight domestic violence?
You can't control everything, and you shouldn't want to. Giving a certain small group control over a much larger group is not a good idea, because you can never know that that small group will handle their power responsibly.
And domestic violence and crime happening on messaging platforms can still be dealt with in the traditional way: through our court system. And that happens and it works and it is fair (at least in essence, not counting corruption).
Stop the CSAM where it’s being created. Catch the drug manufacturers and distributors. Etc. do police work.
I think you are 100% right that above-the-law communication is not good for society. This should be obvious. At the same time, allowing government to be able to spy everywhere is also not good for society. Also obvious. The correct solution is therefore something in the middle.
I'm not convinced by the other arguments here that usually contains a hint of slippery-slope, what-aboutism or false dichotomy fallacies.
The proposed legislation is terrible: it is not balanced and does not contain any safeguard to avoid abuse. However, it does not mean that the equally terrible situation of having easy way for criminals to avoid justice is the good solution either.
Personally, I think that a good system should be a distributed system where several independent justice organizations share the set of key needed to decrypt (for example, a message can be decrypted only if Amnesty International, Interpol and the Austrian Justice Department put their 3 keys together, each individual key being useless on its own). In this model, abuses are almost impossible while obvious crime can still be investigated. I don't know any argument that really works to say that this model is not always better than the free-for-all-all-anonymised-messaging.
Such ideas already exist, and David Chaum even came up with proof of concept of something similar https://www.wired.com/2016/01/david-chaum-father-of-online-a...
That conclusion is logically wrong and does not follow from the premises. You are decrying fallacies in other arguments while making them yourself.
https://en.wikipedia.org/wiki/Argument_to_moderation
> The proposed legislation is terrible: it is not balanced and does not contain any safeguard to avoid abuse.
Then the logical course of action is to try to stop this legislation. Everything else about the argument is irrelevant right now, what matters is the close very bad thing in front of us that we can do something about. Eliminate that and then we can have a reasoned discussion about what the proper approach is.
> several independent justice organizations share the set of key needed to decrypt
There’s no such thing as keys that only good guys have access to. It has been shown time and again that someone with access will abuse it or be tricked.
I won’t go long on this point, however, as I was not familiar with your specific example. I’ll read up more on it. But again, that’s a conversation that matters later.
The correct solution should be something in the middle. Old-fashioned wiretapping, with a warrant and the need to dedicate staff to installing and monitoring the tap is basically okay. The problem is that the mathematics of cryptography and the scaling inherent to information technology mean that only all-or-nothing solutions are possible. If the cryptography is intentionally broken, it's broken not just for law enforcement, it's broken for everyone. If law enforcement has a backdoor they can use with a warrant, they're capable of using it without a warrant, and probably will. And if their special keys get leaked, then again, the encryption is broken for everyone.
Like you point out, secret sharing is one way of getting around this in principle. But governments would never make their access dependent on an NGO; in practice I'm sure they'd only agree to secret sharing schemes where the separate parties were separated only by nominal bureaucratic firewalls, and then you're back to the original problem.
> If we don’t hijack privacy in speech, how do we fight crime happening in private conversations? If [the] government doesn’t have access to what you say at home, what’s stopping criminals from using their homes and never getting tracked down? Or proven guilty, since all the proof was said behind closed doors? Aren’t we hurting ourselves by being so obsessed with privacy?
Should we be obsessed with privacy, or should we let the government put microphones in every house just in case there are paedophiles talking about their sex acts and hence getting away with it?
Similar arguments can be made by substituting other things that were traditionally considered the domain of only authoritarian dictatorships, such as opening all letters and reading them before they're delivered by the postal service, or keeping tabs on what books you borrow from the library.
I grew up in one of those countries, and I can tell you that it's not at all nice that they tracked what you photocopied, you know, just in case you wanted to print out anti-party ("one" party!) propaganda... I mean... something... something distributing child porn. Yeah, that's it. That's the reason.
That's all photocopiers & printers though, with US manufacturers being ground zero for tracking, IIRC - https://www.instructables.com/Yellow-Dots-of-Mystery-Is-Your...
Same as with DRMs and those annoying "piracy is a crime" banner we got for minutes before watching a DVD. Nobody pirating content has ever seen those.
Terrorists on the other hand will find illegal ways to overcome the restrictions. It's the same with encryption - (open-source) tools with e2ee are already broadly available, sideloading is available - nothing will stop terrorists/other criminals to just install that and continue doing what they want.
On the other hand this chat control opens a huge area of opportunities for govts to spy on citizens or maybe journalists/other politicians that they dislike. And this is if we assume the system doesn't have bugs that would allow third parties/hackers to break it and get all the info by themselves, or bugs that can trigger a false-positive event.
That's why it's a bad idea. Criminals will find ways to overcome the limit, govs will get new tools for suppression (even if current govs are 'good', what if the next govt is some ultra conservative or radical nationalist, do you think they'll not use these new tools?) and normal ppl are basically left without any privacy
Crime should not be approved of, and "crime fighting crime" shouldn't magically get an exception.
"But crime X is way worse than crime Y fighting it".
Crime should certainly be punished, but you cannot punish someone before you can prove their wrongdoing. And we have a court system for that.
Punishing people before they're proven of wrongdoing is criminal in itself. You should't give a certain group of people allowance to put prison collars on others without them having done any crime. At least not if you want to live in a free society.
You’re hurting yourself more by being too lax with it. Remember that a crime is whatever the law says it is. So if an authoritarian government makes it so criticising them becomes a crime and has access to all your communication, good luck ever breaking that cycle. You can use other examples, like making homosexuality illegal.
These are real examples that real governments (or people with a good chance of being elected) want.
Remember that the ultimate goal of these laws is never to “protect the children”—that’s just the convenient given reason, because how could you be against that—but to exert more control over the populace and cement the position of those in power. Even if the current government employs the technology only for good—highly unlikely—you don’t know about the next one.
If?
https://www.lewik.org/term/15692/defamation-of-persons-in-th...
What you end up is just having the gov having all information about private citizens, which can be used against us.
For example, the government of Spain constantly shares private citizen data of people near political adversaries for political reasons, also, I don't want people knowing what I share with my girlfriend, regardless of it they do something or not with it publicly.
We need to compromise all door locks, in order to keep everyone safe from criminals.
And with AI… ooft. AI will get to a point where it takes over, and decisions like these help it to. We are destroying our future fast.
Who said it is supposed to be the solution?
Almost no crime problems have a the solution. Instead reducing crime is almost always a matter of a variety of measures that each make the crime a little less likely.
Because here it is mostly facebook/messenger, discord, instagram, and a distant whatsapp, in order.
You are urged to take action, but neither of the linked posts makes it easy for people to do that. Ok, there is a link to a website where I can find some email address for "Permanent Representation of <country name>", thanks. Is it it? Should I send an email? Like, do I have to prove I'm a citizen of <country name> or is emailing them anonymously from some batman69@gmail.com email account is totally fine? I have no idea. Who is my counter-agent, how should I talk to them? Should I assume they are very well aware of the proposal and have their opinion, or are they most likely ignorant about all that stuff? What do I write? What this is supposed to achieve? Do I just email the url to this blog-post? Well, this should be actually a preferable solution, I imagine, but the blogpost doesn't seem to be written in a way to be convincing to a random MEP (assuming he NEEDS to be convinced).
I don't claim I could do better (I'm totally ignorant about all that stuff), but just saying it isn't as actionable as it could probably be.
Are there any studies on the effectiveness of this? To me it always read like a joke. The only way I can imagine this going is.
>Good day Mr. Rep I'm really really worried about x legistlation for reasons y and z.
>Sure buddy. hangs up
>Votes like his "donors" asked him to anyways.
https://www.economist.com/britain/2024/04/11/why-most-people...
In the aftermath of Brexit I remember reading time and again of people who voted to leave and by doing so screwed up their own business (e.g. florists whose flowers come from abroad). They expressed nothing but regret.
So no, the UK didn’t vote leave because they “hate the unelected, undemocratic, wannabe-communist institutions”, they voted leave because they didn’t understand the big picture and were tricked by unscrupulous politicians making false claims (probably the most famous being that a lot more money would go into the national heath service).
"So no, the UK didn’t vote leave because they “hate the unelected, undemocratic, wannabe-communist institutions”,"
Yes they very definitely did. The economic arguments were never the main drivers of Brexit. It was about national sovereignty, and electoral accountability - and this new Orwellian EU law, (enacted "for the children", of course) only serves to reinfoce that view.
There is no chat control in EU yet and there will not probably be one as envisioned by the gloomiest commentators, regardless on how the voting will go.
1 - online child abuse it's real, yes, it can be used as a pretext to sponsor unpopular laws, but it exists and can't be dismissed only as a false claim or something politicians bring up for their evil plan of global dominance and massive surveillance (last sentence is the rhetorical equivalent of What about the children?)
2 - political groups laser focusing on how the World will end if the EU discuss about some issue that has privacy implications aren't in any way better than those laser focusing on childrens' safety.
3 - the elected representatives are elected by the people of EU through a pure proportional electoral system, they represent the Europeans in almost perfect proportion to the population. The issues they are focused on are the issues that people of EU care about, we might not like the opinions and/or the acts of the majority, but it's how democracy work.
4 - the parliament is not new, the new elected representatives are still forming the new political groups (starting from today) so, no, there's no evil puppet master behind the regular schedule of the parliament
5 - issues like this one have been discussed in Europe for 15 years at least, it's nothing new, but every time there is someone predicting the end of the World as we know it. Why? it's simple: it's their job, they've been elected by a tiny minority, have little or no space compared to the larger political groups, so they need to be very local about it and use the sharpest tool of them all: fear. It's political marketing 101, nothing to see here, please disperse naked gun's style. In this case threema makes, not surprisingly, a chat app.
6 - Moreover, in November 2023 the parliament sided with encryption, by not approving the proposal to break E2E encryption while in March this year (2024) The European Court of Human Rights ruled that weakening encryption can violate fundamentals human rights (link to the sentence: https://hudoc.echr.coe.int/eng/#{%22itemid%22:[%22001-230854...}). Democracy also means trusting the system, even when it seems that everything is lost.
7 - the law hasn't passed yet and the voting will probably be delayed again, due to the aforementioned parliament still not actually in place at full steam.
8 - if being European taught me something is that approving a law is one thing, enforcing it is a complete different story. In many countries, like mine, Italy, law enforcement is a farce, especially when it's about regulatory infringements.
Last but not least, laws protecting children have already been abused here in Italy (and I'm sure everywhere else) long before internet was a thing or without using any form of chat control just for political gain.
A couple of examples (sorry, Italian only)
https://it.wikipedia.org/wiki/Diavoli_della_Bassa_modenese
https://www.ilpost.it/2023/06/08/caso-bibbiano/
the second one has been so controversial (and the accusations in such bad faith) that the equivalent Wikipedia page has been removed and locked by the admins https://it.wikipedia.org/wiki/Angeli_e_demoni_(inchiesta).
That is because the people who need to leave office will usually join with the opposition to get laws passed that they wanted but were afraid to vote for before the election.
https://www.europarl.europa.eu/plenary/en/votes.html?tab=ord...
https://www.youtube.com/watch?v=hzYOdO4pEyI
It works in real life too. Distract the public for long enough that few people make a stink and the law gets through. When people complain later it’s “Oops, we didn’t know, no one seemed to care. Well, nothing we can do now”. Much harder to do that if everyone is shouting at you to not do the thing.
It really feels like a symptomatic phenomenon of our time.
Quite often a government body has missed some performance targets, suffered cost overruns or has other bad news which they need to announce publicly at some point. But they can choose when the announcement comes out.
Then along comes September 11th 2001, planes crash into the twin towers, and while the towers are still burning government PR teams are rushing out the announcement that they've badly missed their train punctuality targets.
They know the news and social media are going to be full of the big event for days or weeks. By the time things are quiet enough that the newspapers have space to report on train punctuality, the bad figures are old news.
This works equally well with big good-news stories like royal weddings and big sporting events.
The "good day to bury bad news" quote is interesting because someone leaked an e-mail where a government PR boss literally encouraged it. Usually such encouragement would be by telephone or whatsapp to avoid creating a paper trail.
At least in France, the upcoming Olympics are a strong contender. That and the surprise parliament election our president dropped on us; because he apparently didn't like the result his party got at the latest EU election, but honestly I don't see how he imagine he's going to get a better result this time around.
https://www.europarl.europa.eu/plenary/en/votes.html?tab=ord...
... the submitted article is complete nonsense "EU citizens would no longer be able to communicate in a safe and private manner on the Internet." ..
no, here's the draft law
https://cdn.netzpolitik.org/wp-upload/2024/05/2024-05-28_Cou...
see page 39,
"Without prejudice to Article 10a, this Regulation shall not prohibit or make impossible end-to-end encryption, implemented by the relevant information society services or by the users."
It's a broad framework and - based on my cursory reading:
- providers have to set up a counter-abuse team and fund it
- authorities and industry-wide cooperation on trying to come up with guidelines and tech
- counter-abuse team needs to interpret the guidelines, do "due diligence"
- provider needs to have monitoring to at least have an idea of abuse risks
- if there are, work on addressing them if possible without breaking privacy
As far as I understand the point is have more of services like "YouTube for Kids", where you can give your kid an account and they can only see stuff tagged "kid appropriate" (and YT simply said we are going to be sure there are no bad comments, so there's no comment section for these videos - which hurts their engagement, which hurts profitability).There's a section about penalties and fines, up to 6% of global revenue, if the provider doesn't take abuse seriously. And - again, based on my understanding - this is exactly to prod big services to make these "safer, but less profitable" options.
see page 45 for actual things providers might need to implement
Those advocating for it have a visceral pitch: your kids will be in danger if this doesn't pass. Those against are arguing for abstracts, like freedom and privacy. And sure, people should be in favour of those things - but there's a visceral pitch against, and it's not being used
What will happen if this stuff is put in place, and used as claimed? There will be a massive rush of automated "accusations" of child abuse against innocent parents, grandparents, and other relatives. This will result in many families being unnecessarily stressed and disrupted by investigations, and at least some children being removed from innocent families.
Why will this happen? after all, won't the scanning be 100% accurate?</s>
Obviously not. Tech companies don't want to be responsible for this - the scanning won't be optimised for accuracy, it will be optimised to pass the buck as hard as possible[1], because the tech companies don't want to business of taking the blame when some pedophile isn't caught.
Who is the second level of review? Maybe some minimum wage zero-hours subcontractor at Serco or Group 4 Security, or whoever the equivalent of that is in the EU. But that's probably it - once the image has been labelled 'bad' no-one else will want to look at it.
So, even if you are sending a picture of your own child to your own mother, you will have at the very least to have to think about whether it could be mistaken for child abuse by someone you don't know who has about 5 seconds to look at it and is probably from a completely different culture.
Yet none of this has been brought up in the media either by tech companies or by privacy activists.
[1] The other option is that the tech companies abdicate completely and just use some black box from the government to scan every picture. Problems with that left as an exercise...
Courts convict innocent people all the time, hence no irreversible punishments.
When proposing “permanent” punishments like this, always put yourself in the shoes of someone who is falsely convicted. And consider that false convictions can (and do) happen for a wide variety of reasons: racial bias, political bias, cover ups, government oppression, etc.
More likely, the true goal is to introduce novel forms of policing, and widen them bit by bit until we do become true surveillance states, where every word any citizen ever said is stored in some government archive.
That's orthogonal to whether or not we should implement a means to catch those that transmit child pornography.
Is castration intended to "cure" the criminal or do you think it would act as a deterrent? Would you consider castration for any kind of rape?
It sounds as though the next obvious step would be to amputate the hand of the serial shoplifter, etc. I'm personally not in favor of giving the state that kind of irreversible authority over our body parts.
Because castrating people as punishment ‘because’ they were abused as children doesn’t feel right?
Of course, doing that as part of treatment could in extreme situations be justified, but luckily there’s ‘reversible castration’ (https://en.wikipedia.org/wiki/Chemical_castration), and that (of course) is preferred as treatment, if such harsh measures are required.
What's the alternative? Communication tools that don't have a central server. What's the problem here? This legislation has the potential to unlock a golden age for local-first software
However I have an unpopular opinion, interested to hear what others might think:
We should eliminate anonymity online. If you go on the internet everything you do should be tied back to your name. This can be done using device attestation. Everyone gets a private key tied to their name/address.
This is compatible with free speech. In fact it promotes free speech because being a "troll" becomes a lot more personal.
I think this way of living would be closer to our nature as tribal primates. It would improve behaviour and overall quality of life. Our brains are designed to have checks and balances from wider society which you don't get anonymously online.
This would also reduce the need for govt monitoring because any chat online could be "turned in" by an informer and then any criminals identified.
https://support.apple.com/en-gb/guide/deployment/dep28afbde6...
To do TOR you would need to look at the IP protocol and signing at the packet level. Definitely more difficult.
This is where you completely break your premise, this is Stasi levels of informing, asking for the population to spy on each other. It's not healthy to society when you feel that any other person you interact with might be informing on you to the State, you leave a very wide avenue open for misuse when the State changes its rulings on what's considered criminal.
How is it different from now when you can record a conversation you have? or forward an email to the police?
Unless you're careful and quite tech literate this is already the case
1. Who is allowed to detect, process or store your identity? For what purposes?
2. What about "right to forget"? Does the data need to be destroyed after a certain period?
3. Who manufacturers and sells these private keys? What happens if I lose mine, or it gets stolen?
4. How does this work internationally? Can a key from China access a system in the US?
1. Give my name and address to activate a device 2. "The internet" requires authentication via the HSM.
for certain platforms. IMO platforms should be able to decide for themselves whether they want the option to have people verify themselves via ID or not.
It's the government's job to provide this service
All in all I would be willing to be quarantined if that meant the bots would suddenly die.