diff --git a/en/bigbrother.html b/en/bigbrother.html index 9b2ebce..134ccff 100644 --- a/en/bigbrother.html +++ b/en/bigbrother.html @@ -46,61 +46,61 @@ Prevention is at the heart of my proposal, but prevention alone will never be en In a recent case the dutch police arrested a man. He has been raping his baby daughter. The police could stop these rapes, rescue this little girl and two other children, because the man got detected their views in facebook messenger. -Prevention did not stop these rapes.🗨 -A baby cannot say no or call a hotline. +Prevention did not stop these rapes. +A baby cannot say no or call a hotline.🗨 My proposal does not introduce detection in online messages. Companies have been detecting child sexual abuse for more than 10 years in online messaging.🗨 Last year they sent 1.5 million reports and reviews from the european union alone. -Containing more than 5 million videos and photos of sexual violence against children. +Containing more than 5 million videos and photos of sexual violence against children.🗨 [inaudible] One million of these reports, the majority, came from online messages, e-mail and chat. Every single one could contain the vital clue. -The often only clue that could lead to an investigation that rescues the child. +The often only clue that could lead to an investigation that rescues the child.🗨 Next year, this detection for child sexual abuse in online messages will be forbidden. Totally in the EU. -Unless there is a specific law that allows it. +Unless there is a specific law that allows it.🗨 My proposal restricts detection compared to [Ursula von der Leyen?]. But my proposal makes sure that it is not totally stopped. Allowing detection of the views to continue. -Only when absolutely necessary. +Only when absolutely necessary.🗨 With new safeguards that are regulating big tech companies right now. Who, the big companies today detect, when they like, what they like and how they like (more or less). -My proposal will limit the detection in time, in place and by law. +My proposal will limit the detection in time, in place and by law.🗨 The only offer for risk assessment. Only if there is a significant risk of child sexual abuse. -Only of the prevention on mitigative measures that failed or was not enough. +Only of the prevention on mitigative measures that failed or was not enough.🗨 Only of through court decisions. -Only with approved technology. -Only based on verified indicators of child sexual abuse. +Only with approved technology.🗨 +Only based on verified indicators of child sexual abuse.🗨 Only then they will be allowed to do the detection. None of these safeguards exist today. That means we will have new safeguards compared with the current situation with my proposal. But there is one big difference. Today it is voluntary for the companies to detect. In my proposal this process will also be obligatory. -When companies are allowed to detect, they also must detect. +When companies are allowed to detect, they also must detect.🗨 And if [???] it only detects the views. It is like a police dog at the airport. Sniffing for drugs offer high risk price. Officers open only suspicious packages when the dog barks. -And like police dogs, artifical intelligence must always be subject to human oversight. +And like police dogs, artifical intelligence must always be subject to human oversight.🗨 And the new EU center I am proposing will be a stake for and filter reports. -That no false positives reach the police. +That no false positives reach the police.🗨 And I will not excempt encrypted services if there is a significant risk of child sexual abuse. -Today, detection for malware is allowed and possible and carried out in encrypted environments. +Today, detection for malware is allowed and possible and carried out in encrypted environments.🗨 This is happening today. -And I think that we must protect our children as least as much as we protect our devices from malware. +And I think that we must protect our children as least as much as we protect our devices from malware.🗨 I know you all feel very strongly about this. You also asked me to withdraw my proposal. I will not do that. Because if I should, then from next year, all detection in messages that has been happening for 10 years will be forbidden in the EU. Forbbiden in the Netherlands. -Detection will stop. Rapes will continue undetected. And that little baby I told you about: Her father would probably still be raping her. +Detection will stop. Rapes will continue undetected. And that little baby I told you about: Her father would probably still be raping her.🗨 And what about the crimes will say about the victims? That the victims, of that baby [???] Amsterdam, who confessed to abusing and raping 90 babies and small children. Eight years later, videos are still out there online. The worst moment of the child's life shared perpertual. -Drama upon drama. +Drama upon drama.🗨 The european union is a global leader in protecting online privacy. I am proud of that. We should all be proud of that. -I also want us to become a global leader in protecting children. +I also want us to become a global leader in protecting children.🗨 For one of the worst crimes that we can ever imagine. I'm proud of our new standards of privacy protection. But we must also protect the privacy of the victims. @@ -155,11 +155,167 @@ And if a platform provider learns that a certain URL points to illegal content,
-Chatcontrol however, is a new type of surveillance, that is highly pervasive and that invades personal online spaces in a way that has not been seen to this extend before. The fact that facebook is scanning personal messages does not really say anything, because facebook has always been a terrible company when it comes to privacy. +Chatcontrol however, is a new type of surveillance, that is highly pervasive and +that invades personal online spaces in a way that has not been seen to this extent before.
-(to be continued)
++This is a real problem and it needs to be addressed. But the methods to address it need to be reasonable. +
+ ++This line of thinking is exactly what leads to mass surveillance. +While it is true that real clues can be detected with this method, +it is to be expected that the sheer volume of false positives will make it less likely that the relevant reports are processed. +The authorities who process these reports today are already overwhelmed. +Furthermore, it is likely that the intimate life of millions of innocent people will be violated due to this surveillance. + +
+An approach that focusses on quality instead of quantity is needed. +
+ ++The platforms will continue to have the option and duty to delete illegal content that they are aware of even without chatcontrol. +This is true for both, the old chatcontrol directive and the new chatcontrol proposal. +Platform providers are and will still be able to use their own terms and conditions to allow them to scan data on their platform as they see fit. +
+ ++Not quite. Most communication software has the potential to be used by children. +And children today are using it to communicate with friends, family, teachers, trainers and more. +And the same software is used by adults to fulfill their needs to communicate. +Chatcontrol subjects many innocent citizens to surveillance, including the children! +
+ ++Is that really the EU we want our children to grow up in? +An EU where all communication of children is scrutinized? +Children also have a right and need to privacy! +
+ ++The chatcontrol law that is currently and temporarily in effect does not define as precisely where the scanning needs to take place, +because scanning is voluntary and service and platform providers can themselves decide if they want to scan. +The new chatcontrol law makes scanning mandatory for a lot of services and platforms. +The new proposal will therefore drastically increase surveillance in the EU. +
+ ++This scanning will still affect all users of a legally provided service or software, even if it is only abused by a minority of users. +
+ ++The fact that this technology is approved by the +"EU centre to prevent and combat child sexual abuse", +does not provide any meaningful assurance. +Similarily, the database of indicates that the EU centre manages, is intransparent by design. +
+ ++This is inaccurate. According to Article 44.1 there are 3 different types of indicators: +
+This forces companies to surveil their customers/users. +Most companies view their customers/users with respect +and are probably not happy about this requirement. +
+ + ++The kind of information that is revealed through chatcontrol is vastly more invasive +than any baggage search at the airport. +
+ ++That is besides the point. +Private conversations should enjoy the respect that they deserve and only be available to the intendended recipients. +That means no mass surveillance and no circumvention of encryption. +
+ ++No, it does not. Malware needs to be unencrypted to be executable. +Encrypted malware looks like random data and can not be reasonably detected. +
+ ++Malware needs to be executable to be harmful. +
+ ++Malware detection works on a different axis than child abuse detection. +Personal data is therefore rarely processed by a malware scanner. +
+ ++Yes. And that would be a reasonable outcome. +Investigators have other tools to detect and fight child sexual abuse. +It is technically possible to circumvent the client-side scanning that chatcontrol imposes. +And these methods of circumvention will be adopted by the offenders sooner or later. +Thus, it is mostly innocent citizens who will be subjected to the surveillance of chatcontrol. +Client-side scanning is a terrible idea. +
+ ++Unfortunately, yes. +But it is unreasonable to expect chatcontrol to solve this. +
+ ++A noble goal. +But mass surveillance is the wrong approach. +