update bigbrother page

This commit is contained in:
om 2023-02-26 12:59:25 +01:00
parent 57ae2e8c4d
commit 81a49e6633
1 changed files with 176 additions and 20 deletions

View File

@ -46,61 +46,61 @@ Prevention is at the heart of my proposal, but prevention alone will never be en
In a recent case the dutch police arrested a man.
He has been raping his baby daughter.
The police could stop these rapes, rescue this little girl and two other children, because the man got detected their views in facebook messenger.
Prevention did not stop these rapes.<a href="#baby_prevention">🗨</a>
A baby cannot say no or call a hotline.
Prevention did not stop these rapes.
A baby cannot say no or call a hotline.<a href="#baby_prevention">🗨</a>
My proposal does not introduce detection in online messages.
Companies have been detecting child sexual abuse for more than 10 years in online messaging.<a href="#not_introducing">🗨</a>
Last year they sent 1.5 million reports and reviews from the european union alone.
Containing more than 5 million videos and photos of sexual violence against children.
Containing more than 5 million videos and photos of sexual violence against children.<a href="#magnitude">🗨</a>
[inaudible]
One million of these reports, the majority, came from online messages, e-mail and chat.
Every single one could contain the vital clue.
The often only clue that could lead to an investigation that rescues the child.
The often only clue that could lead to an investigation that rescues the child.<a href="#only_clue">🗨</a>
Next year, this detection for child sexual abuse in online messages will be forbidden.
Totally in the EU.
Unless there is a specific law that allows it.
Unless there is a specific law that allows it.<a href="#expired_law">🗨</a>
My proposal restricts detection compared to [Ursula von der Leyen?].
But my proposal makes sure that it is not totally stopped.
Allowing detection of the views to continue.
Only when absolutely necessary.
Only when absolutely necessary.<a href="#absolutely_necessary">🗨</a>
With new safeguards that are regulating big tech companies right now.
Who, the big companies today detect, when they like, what they like and how they like (more or less).
My proposal will limit the detection in time, in place and by law.
My proposal will limit the detection in time, in place and by law.<a href="#limit_by_law">🗨</a>
The only offer for risk assessment. Only if there is a significant risk of child sexual abuse.
Only of the prevention on mitigative measures that failed or was not enough.
Only of the prevention on mitigative measures that failed or was not enough.<a href="#prevention_not_enough">🗨</a>
Only of through court decisions.
Only with approved technology.
Only based on verified indicators of child sexual abuse.
Only with approved technology.<a href="#approved_technology">🗨</a>
Only based on verified indicators of child sexual abuse.<a href="#only_verified_abuse">🗨</a>
Only then they will be allowed to do the detection.
None of these safeguards exist today.
That means we will have new safeguards compared with the current situation with my proposal.
But there is one big difference.
Today it is voluntary for the companies to detect.
In my proposal this process will also be obligatory.
When companies are allowed to detect, they also must detect.
When companies are allowed to detect, they also must detect.<a href="#allowed_to_detecd">🗨</a>
And if [???] it only detects the views.
It is like a police dog at the airport.
Sniffing for drugs offer high risk price.
Officers open only suspicious packages when the dog barks.
And like police dogs, artifical intelligence must always be subject to human oversight.
And like police dogs, artifical intelligence must always be subject to human oversight.<a href="#dog">🗨</a>
And the new EU center I am proposing will be a stake for and filter reports.
That no false positives reach the police.
That no false positives reach the police.<a href="#no_false_positive_to_police">🗨</a>
And I will not excempt encrypted services if there is a significant risk of child sexual abuse.
Today, detection for malware is allowed and possible and carried out in encrypted environments.
Today, detection for malware is allowed and possible and carried out in encrypted environments.<a href="#encrypted_environments">🗨</a>
This is happening today.
And I think that we must protect our children as least as much as we protect our devices from malware.
And I think that we must protect our children as least as much as we protect our devices from malware.<a href="#malware">🗨</a>
I know you all feel very strongly about this.
You also asked me to withdraw my proposal. I will not do that.
Because if I should, then from next year, all detection in messages that has been happening for 10 years will be forbidden in the EU. Forbbiden in the Netherlands.
Detection will stop. Rapes will continue undetected. And that little baby I told you about: Her father would probably still be raping her.
Detection will stop. Rapes will continue undetected. And that little baby I told you about: Her father would probably still be raping her.<a href="#expired_legislation">🗨</a>
And what about the crimes will say about the victims? That the victims, of that baby [???] Amsterdam, who confessed to abusing and raping 90 babies and small children.
Eight years later, videos are still out there online.
The worst moment of the child's life shared perpertual.
Drama upon drama.
Drama upon drama.<a href="#drama">🗨</a>
The european union is a global leader in protecting online privacy.
I am proud of that.
We should all be proud of that.
I also want us to become a global leader in protecting children.
I also want us to become a global leader in protecting children.<a href="#global_leader">🗨</a>
For one of the worst crimes that we can ever imagine.
I'm proud of our new standards of privacy protection.
But we must also protect the privacy of the victims.
@ -155,11 +155,167 @@ And if a platform provider learns that a certain URL points to illegal content,
</p>
<p>
Chatcontrol however, is a new type of surveillance, that is highly pervasive and that invades personal online spaces in a way that has not been seen to this extend before. The fact that facebook is scanning personal messages does not really say anything, because facebook has always been a terrible company when it comes to privacy.
Chatcontrol however, is a new type of surveillance, that is highly pervasive and
that invades personal online spaces in a way that has not been seen to this extent before.
</p>
<p>(to be continued)</p>
<h2 id="magnitude">A lot of sexual abuse material is shared on the internet</h2>
<p>
This is a real problem and it needs to be addressed. But the methods to address it need to be reasonable.
</p>
<h2 id="only_clue">Images that are shared online can be the only clue for sexual abuse</h2>
<p>
This line of thinking is exactly what leads to mass surveillance.
While it is true that real clues can be detected with this method,
it is to be expected that the sheer volume of false positives will make it less likely that the relevant reports are processed.
The authorities who process these reports today <a href="https://www.bitsoffreedom.nl/2022/10/17/sex-crimes-unit-already-overwhelmed-and-eu-lawmakers-will-only-make-it-worse/">are already overwhelmed</a>.
Furthermore, it is likely that the intimate life of millions of innocent people will be violated due to this surveillance.
<p>
An approach that focusses on quality instead of quantity is needed.
</p>
<h2 id="expired_law">Without the updated legislation it will become illegal again to scan user's data</h2>
<p>
The platforms will continue to have the option and duty to delete illegal content that they are aware of even without chatcontrol.
This is true for both, the old chatcontrol directive and the new chatcontrol proposal.
Platform providers are and will still be able to use their own terms and conditions to allow them to scan data on their platform as they see fit.
</p>
<h2 id="absolutely_necessary">Scanning will only happen when absolutely necessary</h2>
<p>
Not quite. Most communication software has the potential to be used by children.
And children today are using it to communicate with friends, family, teachers, trainers and more.
And the same software is used by adults to fulfill their needs to communicate.
Chatcontrol subjects many innocent citizens to surveillance, including the children!
</p>
<p>
Is that really the EU we want our children to grow up in?
An EU where all communication of children is scrutinized?
Children also have a right and need to privacy!
</p>
<h2 id="limit_by_law">The new chatcontrol law is a lot more targeted than its predecessor</h2>
<p>
The chatcontrol law that is currently and temporarily in effect does not define as precisely where the scanning needs to take place,
because scanning is voluntary and service and platform providers can themselves decide if they want to scan.
The new chatcontrol law makes scanning mandatory for a lot of services and platforms.
The new proposal will therefore drastically increase surveillance in the EU.
</p>
<h2 id="prevention_not_enough">Scanning will only happen if mitigations failed or were not enough</h2>
<p>
This scanning will still affect all users of a legally provided service or software, even if it is only abused by a minority of users.
</p>
<h2 id="approved_technology">Scanning happens only with approved technology</h2>
<p>
The fact that this technology is approved by the
"EU centre to prevent and combat child sexual abuse",
does not provide any meaningful assurance.
Similarily, the database of indicates that the EU centre manages, is intransparent by design.
</p>
<h2 id="only_verified_abuse">Detection will only happen based on verified indicators of child sexual abuse</h2>
<p>
This is inaccurate. According to Article 44.1 there are 3 different types of indicators:
<ol>
<li>Indicators based on previously detected and identified abuse material</li>
<li>Indicators NOT based on previously detected and identified abuse material.</li>
<li>Indicators for detecting of illegal child solicitation</li>
</ol>
And these indicators must be used by hosting and communication providers as described in Article 10.1.
</p>
<h2 id="allowed_to_detecd">If companies are allowed to detect, they must detect</h2>
<p>
This forces companies to surveil their customers/users.
Most companies view their customers/users with respect
and are probably not happy about this requirement.
</p>
<h2 id="dog">Chatcontrol is no worse than a police dog at the airport</h2>
<ul>
<li>The airport is not a private living space.</li>
<li>The airport is not where the police routinely looks into your private communications.</li>
<li>The airport is not a place that most people tie their identity and innermost feelings to.</li>
</ul>
<p>
The kind of information that is revealed through chatcontrol is vastly more invasive
than any baggage search at the airport.
</p>
<h2 id="no_false_positive_to_police">No false positive will be passed to the police</h2>
<p>
That is besides the point.
Private conversations should enjoy the respect that they deserve and only be available to the intendended recipients.
That means no mass surveillance and no circumvention of encryption.
</p>
<h2 id="encrypted_environments">Scanning for malware works also in encrypted environments</h2>
<p>
No, it does not. Malware needs to be unencrypted to be executable.
Encrypted malware looks like random data and can not be reasonably detected.
</p>
<h2 id="malware">Scanning of private content happens already today to fight malware</h2>
<p>
Malware needs to be executable to be harmful.
</p>
<ul>
<li>Text messages are not executable.</li>
<li>Most images do not contain executable code.</li>
<li>Videos do not contain executable code.</li>
</ul>
<p>
Malware detection works on a different axis than child abuse detection.
Personal data is therefore rarely processed by a malware scanner.
</p>
<h2 id="expired_legislation">If we do not accept the new proposal, the scanning against child abuse can not continue</h2>
<p>
Yes. And that would be a reasonable outcome.
Investigators have other tools to detect and fight child sexual abuse.
It is technically possible to circumvent the client-side scanning that chatcontrol imposes.
And these methods of circumvention will be adopted by the offenders sooner or later.
Thus, it is mostly innocent citizens who will be subjected to the surveillance of chatcontrol.
Client-side scanning is a terrible idea.
</p>
<h2 id="drama">CSAM will stay on the internet perpetually</h2>
<p>
Unfortunately, yes.
But it is unreasonable to expect chatcontrol to solve this.
</p>
<h2 id="global_leader">We want to become a global leader in protecting children</h2>
<p>
A noble goal.
But mass surveillance is the wrong approach.
</p>
</div>
</body>