BETA

30 Amendments of Morten LØKKEGAARD related to 2022/0155(COD)

Amendment 80 #
Proposal for a regulation
Recital 70 a (new)
(70 a) In line with Directive 2011/93/EU of the European Parliament and of the Council, this Regulation recognises and safeguards the key role of hotlines in order to enhance the fight against child sexual abuse online in the European Union. Hotlines have a track-record of proven capability since 1999 in the identification and removal of child sexual abuse material from the digital environment and have created a worldwide network and procedures for the child sexual abuse identification and removal. Member States should therefore promote and safeguard the role of formally recognized non-governmental organizations involved in anonymous public reporting of child sexual abuse material, which are at the forefront of detecting new child sexual abuse material, which is an essential factor in finding new victims while also keeping the databases of indicators up to date.
2022/11/30
Committee: CULT
Amendment 93 #
Proposal for a regulation
Article 2 – paragraph 1 – point w a (new)
(w a) ‘hotline’ means an organisation providing a mechanism, other than the reporting channels provided by law enforcement agencies, for receiving anonymous information from the public about alleged child sexual abuse material and online child sexual exploitation, which meets all the following criteria: (a) is officially recognised by its home Member State as expressed in the Directive 2011/93/EU of the European Parliament and of the Council; (b) has the mission of combatting child sexual abuse material in its articles of association; and (c) is part of a recognised and well-established international network of hotlines as referred to in this article.
2022/11/30
Committee: CULT
Amendment 95 #
Proposal for a regulation
Chapter I a (new)
Ia PREVENTION AND EDUCATION PROGRAMMES Article 2 a (new) 1. Member States shall take appropriate measures, such as education, awareness raising campaigns and training, to discourage and reduce the demand that fosters all forms of sexual exploitation of children in the online environment. 2. Member States shall take appropriate action, including through the Internet, such as information and awareness- raising campaigns, research and early- education programmes, where appropriate in cooperation with relevant civil society organisations acting in the public interest against child sexual abuse, law enforcement authorities and other stakeholders, aimed at raising awareness and reducing the risk of children becoming victims of sexual abuse or of exploitation online. 3. Member States shall promote regular training for officials likely to come into contact with child victims of sexual abuse or exploitation online, including the solicitation of children, aimed at enabling them to identify and deal with child victims and potential child victims. 4. Member States shall promote regular training for officials to inform them and update their knowledge on the latest trends in the creation, dissemination and monetization of child sexual abuse materials and national data hosting of child sexual abuse material.
2022/11/30
Committee: CULT
Amendment 98 #
Proposal for a regulation
Article 4 – paragraph 1 – point b a (new)
(b a) to provide, through appropriate technical and operational measures, readily accessible and easy-to-use parental tools to help parents or guardians support children and identify harmful behaviour;
2022/11/30
Committee: CULT
Amendment 99 #
Proposal for a regulation
Article 4 – paragraph 1 – point c
(c) initiating or adjusting cooperation, in accordance with competition law, with other providers of hosting services or providers of interpersonal communication services, public authorities, civil society organisations, hotlines or, where applicable, entities awarded the status of trusted flaggers in accordance with Article 19 of Regulation (EU) …/… [on a Single Market For Digital Services (Digital Services Act) and amending Directive 2000/31/EC] .
2022/11/30
Committee: CULT
Amendment 100 #
Proposal for a regulation
Article 6 a (new)
Article 6 a Anonymous public reporting of online child sexual abuse 1. Member States shall take appropriate measures to promote and safeguard the role of formally recognized non- governmental organizations involved in anonymous public reporting of child sexual abuse material and the proactive search for such material. 2. Member States shall ensure that the public always has the possibility to anonymously report child sexual abuse material and child sexual exploitation activities to hotlines specialised in combatting online child sexual abuse material and shall safeguard the role of such hotlines in anonymous public reporting. 3. Member States shall ensure that the hotlines referred to in paragraph 2 operating in their territory are authorised to view, assess and process anonymous reports of child sexual abuse material. 4. Member States shall grant the hotlines referred to in paragraph 2 the authority to issue content removal notices for confirmed instances of child sexual abuse material. 5. Member States shall authorise the hotlines referred to in paragraph 2 to voluntarily conduct pro-active searching for child sexual abuse material online.
2022/11/30
Committee: CULT
Amendment 213 #
Proposal for a regulation
Recital 26
(26) The measures taken by providers of hosting services and providers of publicly available interpersonal communications services to execute detection orders addressed to them should remain strictly limited to what is specified in this Regulation and in the detection orders issued in accordance with this Regulation. In order to ensure the effectiveness of those measures, allow for tailored solutions, remain technologically neutral, and avoid circumvention of the detection obligations, those measures should be taken regardless of the technologies used by the providers concerned in connection to the provision of their services. Therefore, this Regulation leaves to the provider concerned the choice of the technologies to be operated to comply effectively with detection orders and should not be understood as incentivising or disincentivising the use of any given technology, provided that the technologies and accompanying measures meet the requirements of this Regulation. That includes the use of end-to-end encryption technology, which is an important tool to guarantee the security and confidentiality of the communications of users, including those of childrennotably those of children. Any prohibition of encryption could potentially be abused by malicious third parties. The issuance of a detection order should therefore not lead a provider to prohibit encryption. In full accordance and respect of encryption, this regulation encourages the innovation and constant improvement of technology detecting, reporting, and removing of CSAM in all digital spaces that are misused for the purpose of dissemination of such material. When executing the detection order, providers should take all available safeguard measures to ensure that the technologies employed by them cannot be used by them or their employees for purposes other than compliance with this Regulation, nor by third parties, and thus to avoid undermining the security and confidentiality of the communications of users.
2023/03/09
Committee: IMCO
Amendment 292 #
Proposal for a regulation
Article 3 – paragraph 2 – point b – indent 4 a (new)
- Functionalities enabling scanning for known child sexual abuse material on upload;
2023/03/09
Committee: IMCO
Amendment 294 #
Proposal for a regulation
Article 3 – paragraph 2 – point b – indent 4 b (new)
- Functionalities preventing uploads from the dark web;
2023/03/09
Committee: IMCO
Amendment 304 #
Proposal for a regulation
Article 3 – paragraph 2 – point e – point iii – indent 3 a (new)
- Enabling users to create usernames that contain a representation about, or imply, the user’s age;
2023/03/09
Committee: IMCO
Amendment 306 #
Proposal for a regulation
Article 3 – paragraph 2 – point e – point iii – indent 3 b (new)
- Enabling child users to create usernames that contain location information on child users;
2023/03/09
Committee: IMCO
Amendment 307 #
Proposal for a regulation
Article 3 – paragraph 2 – point e – point iii – indent 3 c (new)
- Enabling users to know or infer the location of child users.
2023/03/09
Committee: IMCO
Amendment 315 #
Proposal for a regulation
Article 3 – paragraph 3 a (new)
3 a. Providers of hosting services and providers of interpersonal communication services shall put forward specific age assurance verification systems that meet the following criteria: (a) effectively protect the privacy of users and do not disclose data gathered for the purposes of age assurance for any other purpose (b) do not collect data that is not strictly necessary for the purposes of age assurance; (c) be proportionate to the risks associated to the product or service that presents a risk of misuse of child sexual abuse; (d) provide appropriate remedies and redress mechanisms for users whose age is wrongly identified.
2023/03/09
Committee: IMCO
Amendment 323 #
Proposal for a regulation
Article 4 – paragraph 1 – introductory part
1. Providers of hosting services and providers of interpersonal communications services shall take reasonable mitigation measures, tailored to the risk identified pursuant to Article 3, to minimise that risk. Such measures shall include, but need not to be limited to, some or all of the following:
2023/03/09
Committee: IMCO
Amendment 337 #
Proposal for a regulation
Article 4 – paragraph 1 – subparagraph 1 (new)
If the risk assessment conducted or updated in accordance with Article 3 identifies that there is a risk of use of the service being used to disseminate, store or make available known child sexual abuse material, reasonable mitigation measures may include measures to detect and remove such material.
2023/03/09
Committee: IMCO
Amendment 346 #
Proposal for a regulation
Article 4 – paragraph 2 – point d a (new)
(d a) Providers of hosting services and providers of interpersonal communications services are encouraged to put in place voluntary measures to detect and report online child sexual abuse for those services that have proven to pose a risk of misuse for child sexual abuse, or in cases there is an imminent risk of misue for child sexual abuse, including for the purpose of the solicitation of children;
2023/03/09
Committee: IMCO
Amendment 373 #
Proposal for a regulation
Article 5 – paragraph 6 a (new)
6 a. Providers of hosting services and providers of interpersonal communications services that qualify as micro (or small) enterprises within the meaning of Article 3 of Directive 2013/34/EU shall transmit a simplified version of the report under paragraph 1 of this Article.
2023/03/09
Committee: IMCO
Amendment 399 #
Proposal for a regulation
Article 7 – paragraph 2 – subparagraph 1
The Coordinating Authority of establishment shall, before requesting the issuance of a detection order, carry out the investigations and assessments necessary to determine whether the conditions of paragraph 4 have been met. Detection orders issued by the coordinating authorities shall serve as a measure of last resort, only enacted when all mitigating measures, including voluntary ones, have proven unsuccessful.
2023/03/09
Committee: IMCO
Amendment 425 #
Proposal for a regulation
Article 7 – paragraph 4 – subparagraph 1 – point b a (new)
(b a) The voluntary measures applied as mitigating measures have not proven successful in preventing the misuse of the service for child sexual abuse.
2023/03/09
Committee: IMCO
Amendment 452 #
Proposal for a regulation
Article 7 – paragraph 9 – subparagraph 1
The competent judicial authority or independent administrative authority shall specify in the detection order the period during which it applies, indicating the start date and the end date, within which the providers of hosting services and providers of interpersonal communications services shall prove that their service is no longer used for child sexual abuse.
2023/03/09
Committee: IMCO
Amendment 467 #
Proposal for a regulation
Article 8 – paragraph 1 – point c
(c) the name of the provider and, where applicable, its legal representative, without prejudice to the issuance of detection orders where the legal name of the provider is not readily ascertained;
2023/03/09
Committee: IMCO
Amendment 515 #
Proposal for a regulation
Article 10 – paragraph 4 – point a
(a) take all the necessary and proportionate measures to ensure that the technologies and indicators, as well as the processing of personal data and other data in connection thereto, are used for the sole purpose of detecting the dissemination of known or new child sexual abuse material or the solicitation of children, as applicable, insofar as strictly limited to what is necessary to execute the detection orders addressed to them;
2023/03/09
Committee: IMCO
Amendment 519 #
Proposal for a regulation
Article 10 – paragraph 4 – point c a (new)
(c a) ensure privacy and safety by design and by default and, where applicable, the protection of encryption.
2023/03/09
Committee: IMCO
Amendment 523 #
Proposal for a regulation
Article 10 – paragraph 4 – point e a (new)
(e a) Ensure safety-by-design tools such as parental controls tool and effective age verification tools.
2023/03/09
Committee: IMCO
Amendment 535 #
Proposal for a regulation
Article 12 – paragraph 1 a (new)
1 a. Where a provider of hosting services or a provider of interpersonal communications services receives a report by the public through, among others, trusted hotline, it shall process and analyse the report in a timely and effective manner as to assess an imminent risk of miuse of the service for child child sexual abuse, without prejudice to the obligation to report to the EU centre pursuant paragraph 1.
2023/03/09
Committee: IMCO
Amendment 539 #
Proposal for a regulation
Article 12 – paragraph 2 a (new)
2 a. The report submitted by the provider pursuant paragrah 2, shall never contain information about the source of the report, especially when this stems from the person to whom the material relates.
2023/03/09
Committee: IMCO
Amendment 543 #
Proposal for a regulation
Article 12 – paragraph 3
3. The provider shall establish and operate an accessible, age-appropriate and user-friendly mechanism that allows users to easily flag to the provider potential online child sexual abuse on the service.
2023/03/09
Committee: IMCO
Amendment 560 #
Proposal for a regulation
Article 14 – paragraph 3 – point b
(b) the name of the provider and, where applicable, of its legal representative, without prejudice to the issuance of removal orders where the legal name of the provider is not readily ascertained;
2023/03/09
Committee: IMCO
Amendment 563 #
Proposal for a regulation
Article 14 – paragraph 3 a (new)
3 a. Providers of hosting services or providers of interpersonal communication services shall be encouraged to extend the effect of the order regarding one or more specific items of material, referred to in paragraph 1, to any provider or services under their control and promptly inform the Coordinating Authority of establishment of this specific measure.
2023/03/09
Committee: IMCO
Amendment 670 #
Proposal for a regulation
Article 85 – paragraph 1
1. By [five years after the entry into force of this Regulation], and every five years thereafter, the Commission shall evaluate this Regulation and submit a report on its application to the European Parliament and the Council. This report shall address in particular the possible use of new technologies for a safe and trusted processing of personal and other data and for the purpose of combating online child sexual abuse and in particular to detect, report and remove online child sexual abuse. The report shall be accompanied, where appropriate, by a legislative proposal.
2023/03/09
Committee: IMCO