BETA

Activities of Laurence FARRENG related to 2020/0361(COD)

Plenary speeches (1)

Digital Services Act (continuation of debate)
2022/01/19
Dossiers: 2020/0361(COD)

Amendments (180)

Amendment 128 #
Proposal for a regulation
Recital 9
(9) This Regulation should complement, yet not affect the application of rules resulting from other acts of Union law regulating certain aspects of the provision of intermediary services, in particular Directive 2000/31/EC, with the exception of those changes introduced by this Regulation, Directive 2010/13/EU of the European Parliament and of the Council as amended,28, and Regulation (EU) …/.. of the European Parliament and of the Council29 – proposed Terrorist Content Online Regulation. Therefore, this Regulation leaves those other acts, which are to be considered lex specialis in relation to the generally applicable framework set out in this Regulation, unaffected. This Regulation should also respect the jurisdiction of the Member States in adopting laws promoting press freedom and media pluralism in addition to cultural and linguistic diversity. However, the rules of this Regulation apply in respect of issues that are not or not fully addressed by those other acts as well as issues on which those other acts leave Member States the possibility of adopting certain measures at national level. _________________ 28 Directive 2010/13/EU of the European Parliament and of the Council of 10 March 2010 on the coordination of certain provisions laid down by law, regulation or administrative action in Member States concerning the provision of audiovisual media services (Audiovisual Media Services Directive) (Text with EEA relevance), OJ L 95, 15.4.2010, p. 1 . 29Regulation (EU) …/.. of the European Parliament and of the Council – proposed Terrorist Content Online Regulation
2021/07/23
Committee: CULT
Amendment 134 #
Proposal for a regulation
Recital 11
(11) It should be clarified that this Regulation is without prejudice to the rules of Union law on copyright and related rights, in particular Directive (EU) 2019/790 on copyright and related rights in the Digital Single Market, which establish specific rules and procedures that should remain unaffected.
2021/07/23
Committee: CULT
Amendment 187 #
Proposal for a regulation
Recital 40
(40) Providers of hosting services play a particularly important role in tackling illegal content online, as they store information provided by and at the request of the recipients of the service and typically give other recipients access thereto, sometimes on a large scale. It is important that all providers of hosting services, regardless of their size, put in place user-friendly notice and action mechanisms that facilitate the notification of specific items of information that the notifying party considers to be illegal content to the provider of hosting services concerned ('notice'), pursuant to which that provider can decide whether or not it agrees with that assessment and wishes to remove or disable access to that content ('action'). Provided the requirements on notices are met, it should be possible for individuals or entities to notify multiple specific items of allegedly illegal content through a single notice. The obligation to put in place notice and action mechanisms should apply, for instance, to file storage and sharing services, web hosting services, advertising servers and paste bins, in as far as they qualify as providers of hosting services covered by this Regulation. Moreover, the notification and action mechanism should be supplemented by 'stay down' provisions, under which providers of hosting services should do everything in their power to prevent the reappearance of content which is identical to other content which they have identified and withdrawn as illegal. The application of this requirement must not lead to a general monitoring obligation.
2021/07/23
Committee: CULT
Amendment 221 #
Proposal for a regulation
Recital 11
(11) It should be clarified that this Regulation is without prejudice to the rules of Union law on copyright and related rights, in particular Directive (EU) 2019/790 on Copyright and Related Rights in Digital Single Market, which establish specific rules and procedures that should remain unaffected.
2021/07/08
Committee: IMCO
Amendment 222 #
Proposal for a regulation
Recital 57
(57) Three categories of systemic risks should be assessed in-depth. A first category concerns the risks associated with the misuse of their service through the dissemination of illegal content, such as the dissemination of child sexual abuse material or illegal hate speech, and the conduct of illegal activities, such as the sale of products, or services prohibited by Union or national law, including counterfeit products or the presentation of content which violates copyright. For example, and without prejudice to the personal responsibility of the recipient of the service of very large online platforms for possible illegality of his or her activity under the applicable law, such dissemination or activities may constitute a significant systematic risk where access to such content may be amplified through accounts with a particularly wide reach. A second category concerns the impact of the service on the exercise of fundamental rights, as protected by the Charter of Fundamental Rights, including the freedom of expression and information, the right to private life, the right to non-discrimination and the rights of the child. Such risks may arise, for example, in relation to the design of the algorithmic systems used by the very large online platform or the misuse of their service through the submission of abusive notices or other methods for silencing speech or hampering competition. A third category of risks concerns the intentional and, oftentimes, coordinated manipulation of the platform’s service, with a foreseeable impact on health, civic discourse, electoral processes, public security and protection of minors, having regard to the need to safeguard public order, protect privacy and fight fraudulent and deceptive commercial practices. Such risks may arise, for example, through the creation of fake accounts, the use of bots, and other automated or partially automated behaviours, which may lead to the rapid and widespread dissemination of information that is illegal content or incompatible with an online platform’s terms and conditions.
2021/07/23
Committee: CULT
Amendment 230 #
Proposal for a regulation
Recital 12
(12) In order to achieve the objective of ensuring a safe, predictable and trusted online environment, for the purpose of this Regulation the concept of “illegal content” should be defined broadly and also covers information relating to illegal content, products, services and activities. In particular, that concept should be understood to refer to information, irrespective of its form, that under the applicable law is either itself illegal, such as illegal hate speech or terrorist content and unlawful discriminatory content, or that relates to activities that are illegal, such as the sharing of images depicting child sexual abuse, unlawful non- consensual sharing of private images, online stalking, the sale of non-compliant or counterfeit products, the non-authorised use of copyright protected material or activities involving infringements of consumer protection law. In this regard, it is immaterial whether the illegality of the information or activity results from Union law or from national law that is consistent with Union law, including the EU Charter on Fundamental Rights and what the precise nature or subject matter is of the law in question.
2021/07/08
Committee: IMCO
Amendment 235 #
Proposal for a regulation
Recital 62
(62) A core part of a very large online platform’s business is the manner in which information is prioritised and presented on its online interface to facilitate and optimise access to information for the recipients of the service. This is done, for example, by algorithmically suggesting, ranking and prioritising information, distinguishing through text or other visual representations, or otherwise curating information provided by recipients. Such recommender systems can have a significant impact on the ability of recipients to retrieve and interact with information online. They also play an important role in the amplification of certain messages, the viral dissemination of information and the stimulation of online behaviour. These recommender systems can also have an impact on consumers' media consumption and cultural practices and might lead to their being enclosed in a bubble without enabling them to discover other content. Consequently, very large online platforms should ensure that recipients are appropriately informed, and can influence the information presented to them. They should clearly present the main parameters for such recommender systems in an easily comprehensible manner to ensure that the recipients understand how information is prioritised for them. They should also ensure that the recipients enjoy alternative options for the main parameters, including options that are not based on profiling of the recipient.
2021/07/23
Committee: CULT
Amendment 239 #
Proposal for a regulation
Recital 13
(13) Considering the particular characteristics of the services concerned and the corresponding need to make the providers thereof subject to certain specific obligations, it is necessary to distinguish, within the broader category of providers of hosting services as defined in this Regulation, the subcategory of online platforms. Online platforms, such as social networks, content-sharing platforms, search engines, livestreaming platforms, messaging services or online marketplaces, should be defined as providers of hosting services that not only store information provided by the recipients of the service at their request, but that also disseminate that information to the public, again at their request. However, in order to avoid imposing overly broad obligations, providers of hosting services should not be considered as online platforms where the dissemination to the public is merely a minor and purely ancillary feature of another service and that feature cannot, for objective technical reasons, be used without that other, principal service, and the integration of that feature is not a means to circumvent the applicability of the rules of this Regulation applicable to online platforms. For example, the comments section in an online newspaper could constitute such a feature, where it is clear that it is ancillary to the main service represented by the publication of news under the editorial responsibility of the publisher.
2021/07/08
Committee: IMCO
Amendment 244 #
Proposal for a regulation
Recital 14
(14) The concept of ‘dissemination to the public’, as used in this Regulation, should entail the making available of information to a large or potentially unlimited number of persons, that is, making the information easily accessible to users in general without further action by the recipient of the service providing the information being required, irrespective of whether those persons actually access the information in question. Accordingly, where access to information requires registration or admission to a user group, such information should only be considered to be publicly available when users seeking to access such information are automatically registered or admitted without human intervention to decide or select the users to whom access is granted. The mere possibility to create groups of users of a given service, including a messaging service should not, in itself, be understood to mean that the information disseminated in that manner is not disseminated to the public. However, the concept should exclude dissemination of information within closed groups consisting of a finlimited number of pre- determined persons taking into account the potential for groups to become tools for wide dissemination of content to the public. Interpersonal communication services, as defined in Directive (EU) 2018/1972 of the European Parliament and of the Council,39 such as emails or private messaging services, fall outside the scope of this Regulation where they do not meet the above criteria for "dissemination to the public". Information should be considered disseminated to the public within the meaning of this Regulation only where that occurs upon the direct request by the recipient of the service that provided the information. File-sharing services and other cloud services fall within the scope of this Regulation, to the extent that such services are used to make the stored information available to the public at the direct request of the content provider. __________________ 39Directive (EU) 2018/1972 of the European Parliament and of the Council of 11 December 2018 establishing the European Electronic Communications Code (Recast), OJ L 321, 17.12.2018, p. 36
2021/07/08
Committee: IMCO
Amendment 258 #
Proposal for a regulation
Article premier – paragraph 5 – point c
(c) Union law on copyright and related rights, in particular Directive (EU) 2019/790 on copyright and related rights in the Digital Single Market;
2021/07/23
Committee: CULT
Amendment 299 #
Proposal for a regulation
Recital 25
(25) In order to create legal certainty and not to discourage activities aimed at detecting, identifying and acting against illegal content that providers of intermediary services may undertake on a voluntary basis, it should be clarified that the mere fact that providers undertake such activities does not lead to the unavailability of the exemptions from liability set out in this Regulation, provided those activities are carried out in good faith and in a diligent manner. In addition, it is appropriate to clarify that the mere fact that those providers take measures, in good faith, to comply with the requirements of Union law, including those set out in this Regulation as regards the implementation of their terms and conditions, should not lead to the unavailability of those exemptions from liability. Therefore, any such activities and measures that a given provider may have taken in order to detect, identify and act against illegal content on a voluntary basis should not be taken into account when determining whether the provider can rely on an exemption from liability, in particular as regards whether the provider provides its service neutrally and can therefore fall within the scope of the relevant provision, without this rule however implying that the provider can necessarily rely thereon.
2021/07/08
Committee: IMCO
Amendment 313 #
Proposal for a regulation
Recital 28
(28) Providers of intermediary services should not be subject to a monitoring obligation with respect to obligations of a general nature. This does not concern monitoring obligations in a specific case and, in particular, does not affect orders by national authorities in accordance with national legislation, in accordance with the conditions established in this Regulation. Nothing in this Regulation should be construed as an imposition of a general monitoring obligation or active fact-finding obligation, or as a general obligation for providers to take proactive measures to relation to illegal content. This should be without prejudice to decisions of Member States to require service providers, who host information provided by users of their service, to apply due diligence measures.
2021/07/08
Committee: IMCO
Amendment 322 #
Proposal for a regulation
Recital 29
(29) Depending on the legal system of each Member State and the field of law at issue, national judicial or administrative authorities may order providers of intermediary services to act against certain specific items of illegal content or to provide certain specific items of information. The national laws in conformity with the Union law, including the EU Charter on Fundamental Rights on the basis of which such orders are issued differ considerably and the orders are increasingly addressed in cross-border situations. In order to ensure that those orders can be complied with in an effective and efficient manner, so that the public authorities concerned can carry out their tasks and the providers are not subject to any disproportionate burdens, without unduly affecting the rights and legitimate interests of any third parties, it is necessary to set certain conditions that those orders should meet and certain complementary requirements relating to thensure the effective processing of those orders.
2021/07/08
Committee: IMCO
Amendment 328 #
Proposal for a regulation
Recital 30
(30) Orders to act against illegal content or to provide information should be issued in compliance with Union law, including the EU Charter on Fundamental Rights and in particular Regulation (EU) 2016/679 and the prohibition of general obligations to monitor information or to actively seek facts or circumstances indicating illegal activity laid down in this Regulation. The competent authorities of Member States should be able to object to the Board orders to act against illegal content, that they consider are in breach of the Union law, including the Charter. The procedure for objection should be simplified and fast-tracked when such orders are issued from an administrative or judicial authority of a Member State that is under an Article 7 procedure for infringement of European values according to Article 2 of TEU. The conditions and requirements laid down in this Regulation which apply to orders to act against illegal content are without prejudice to other Union acts providing for similar systems for acting against specific types of illegal content, such as Regulation (EU) …/…. [proposed Regulation addressing the dissemination of terrorist content online], or Regulation (EU) 2017/2394 that confers specific powers to order the provision of information on Member State consumer law enforcement authorities, whilst the conditions and requirements that apply to orders to provide information are without prejudice to other Union acts providing for similar relevant rules for specific sectors. Those conditions and requirements should be without prejudice to retention and preservation rules under applicable national law, in conformity with Union law and confidentiality requests by law enforcement authorities related to the non- disclosure of information.
2021/07/08
Committee: IMCO
Amendment 333 #
Proposal for a regulation
Recital 31
(31) The territorial scope of such orders to act against illegal content should be clearly set out on the basis of the applicable Union or national law in conformity with the Union law, including the EU Charter on Fundamental Rights enabling the issuance of the order and should not exceed what is strictly necessary to achieve its objectives. In that regard, the national judicial or administrative authority issuing the order should balance the objective that the order seeks to achieve, in accordance with the legal basis enabling its issuance, with the rights and legitimate interests of all third parties that may be affected by the order, in particular their fundamental rights under the Charter. In addition, where the order referring to the specific information may have effects beyond the territory of the Member State of the authority concerned, the authority should assess whether the information at issue is likely to constitute illegal content in other Member States concerned and, where relevant, take account of the relevant rules of national, Union law or international law and the interests of international comity.
2021/07/08
Committee: IMCO
Amendment 339 #
Proposal for a regulation
Recital 33
(33) Orders to act against illegal content and to provide information are subject to the rules safeguarding the competence of the Member State where the service provider addressed is established and laying down possible derogations from that competence in certain cases, set out in Article 3 of Directive 2000/31/EC, only if the conditions of that Article are met. Given that the orders in question relate to specific items of illegal content and information as defined in Union or national law in conformity with the Union law, including the EU Charter on Fundamental Rights, respectively, where they are addressed to providers of intermediary services established in another Member State, they do not in principle restrict those providers’ freedom to provide their services across borders. Therefore, the rules set out in Article 3 of Directive 2000/31/EC, including those regarding the need to justify measures derogating from the competence of the Member State where the service provider is established on certain specified grounds and regarding the notification of such measures, do not apply in respect of those orders.
2021/07/08
Committee: IMCO
Amendment 343 #
Proposal for a regulation
Article 14 – paragraph 6 a (new)
6a. Providers of hosting services shall do all in their power to prevent the reappearance of content which is identical to other content which they have already identified and withdrawn as being illegal. The application of this requirement must not lead to any general monitoring obligation.
2021/07/23
Committee: CULT
Amendment 377 #
Proposal for a regulation
Recital 40
(40) Providers of hosting services play a particularly important role in tackling illegal content online, as they store information provided by and at the request of the recipients of the service and typically give other recipients access thereto, sometimes on a large scale. It is important that all providers of hosting services, regardless of their size, put in place user-friendly notice and action mechanisms that facilitate the notification of specific items of information that the notifying party considers to be illegal content to the provider of hosting services concerned ('notice'), pursuant to which that provider can decide whether or not it agrees with that assessment and wishes to remove or disable access to that content ('action'). Provided the requirements on notices are met, it should be possible for individuals or entities to notify multiple specific items of allegedly illegal content through a single notice. The obligation to put in place notice and action mechanisms should apply, for instance, to file storage and sharing services, web hosting services, advertising servers and paste bins, in as far as they qualify as providers of hosting services covered by this Regulation. Furthermore, the notice and action mechanism should be complemented by ‘stay down’ provisions whereby providers of hosting services should demonstrate their best efforts in order to prevent from reappearing content which is identical to another piece of content that has already been identified and removed by them as illegal. The application of this requirement should not lead to any general monitoring obligation.
2021/07/08
Committee: IMCO
Amendment 394 #
Proposal for a regulation
Recital 42
(42) Where a hosting service provider decides to remove or disable information provided by a recipient of the service, for instance following receipt of a notice or acting on its own initiative, including through the use of automated means, that have been proven to be efficient, proportionate and reliable, that provider should inform the recipient of its decision, the reasons for its decision and the available redress possibilities to contest the decision, in view of the negative consequences that such decisions may have for the recipient, including as regards the exercise of its fundamental right to freedom of expression. That obligation should apply irrespective of the reasons for the decision, in particular whether the action has been taken because the information notified is considered to be illegal content or incompatible with the applicable terms and conditions. Available recourses to challenge the decision of the hosting service provider should always include judicial redress.
2021/07/08
Committee: IMCO
Amendment 423 #
Proposal for a regulation
Recital 47
(47) The misuse of services of online platforms by frequently providing manifestly illegal content or by frequently submitting manifestly unfounded notices or complaints under the mechanisms and systems, respectively, established under this Regulation undermines trust and harms the rights and legitimate interests of the parties concerned. Therefore, there is a need to put in place appropriate and, proportionate and effective safeguards against such misuse. Information should be considered to be manifestly illegal content and notices or complaints should be considered manifestly unfounded where it is evident to a layperson, without any substantive analysis, that the content is illegal respectively that the notices or complaints are unfounded. Under certain conditions, online platforms should temporarily suspend their relevant activities in respect of the person engaged in abusive behaviour. This is without prejudice to the freedom by online platforms to determine their terms and conditions and establish stricter measures in the case of manifestly illegal content related to serious crimes. For reasons of transparency, this possibility should be set out, clearly and in sufficiently detail, in the terms and conditions of the online platforms. Redress should always be open to the decisions taken in this regard by online platforms and they should be subject to oversight by the competent Digital Services Coordinator. The rules of this Regulation on misuse should not prevent online platforms from taking other measures to address the provision of illegal content by recipients of their service or other misuse of their services, in accordance with the applicable Union and national law. Those rules are without prejudice to any possibility to hold the persons engaged in misuse liable, including for damages, provided for in Union or national law.
2021/07/08
Committee: IMCO
Amendment 429 #
Proposal for a regulation
Recital 48
(48) An online platform may in some instances become aware, such as through a notice by a notifying party or through its own voluntary measures, of information relating to certain activity of a recipient of the service, such as the provision of certain types of illegal content, that reasonably justify, having regard to all relevant circumstances of which the online platform is aware, the suspicion that the recipient may have committed, may be committing or is likely to commit a serious criminal offence involving a threat to the life or safety of person, notably when it concerns vulnerable users, such as offences specified in Directive 2011/93/EU of the European Parliament and of the Council44 . In such instances, the online platform should inform without delay the competent law enforcement authorities of such suspicion, providing all relevant information available to it, including where relevant the content in question and an explanation of its suspicion. This Regulation does not provide the legal basis for profiling of recipients of the services with a view to the possible identification of criminal offences by online platforms. Online platforms should also respect other applicable rules of Union or national law for the protection of the rights and freedoms of individuals when informing law enforcement authorities. __________________ 44Directive 2011/93/EU of the European Parliament and of the Council of 13 December 2011 on combating the sexual abuse and sexual exploitation of children and child pornography, and replacing Council Framework Decision 2004/68/JHA (OJ L 335, 17.12.2011, p. 1).
2021/07/08
Committee: IMCO
Amendment 433 #
Proposal for a regulation
Recital 48 a (new)
(48a) Where an online platform becomes aware of any information giving rise to a suspicion that a serious criminal offence involving a threat to the life or safety of persons has taken place, is taking place or is likely to take place, it should remove or disable the content and promptly inform the law enforcement or judicial authorities of the Member State or Member States concerned of its suspicion and provide all available relevant information.
2021/07/08
Committee: IMCO
Amendment 436 #
Proposal for a regulation
Article 29 – paragraph 1
1. Very large online platforms that use recommender systems shall set out in their terms and conditions, in a and on a designated webpage, directly accessible from the online interface of the very large online platforms, in a manner which is clear, accessible and easily comprehensible mannerto all, the main parameters used in their recommender systems, and the optimisation objectives sought by these recommender systems as well as any options for the recipients of the service to modify or influence those main parameters that they may have made available, including at least one option which is not based on profiling, within the meaning of Article 4 (4) of Regulation (EU) 2016/679.
2021/07/23
Committee: CULT
Amendment 437 #
Proposal for a regulation
Recital 49
(49) In order to contribute to a safe, trustworthy and transparent online environment for consumers, as well as for other interested parties such as competing traders and holders of intellectual property rights, and to deter traders from selling products or services in violation of the applicable rules, online platforms allowing consumers to conclude distance contracts with tradermarketplaces should ensure that such traders are traceable. The trader should therefore be required to provide certain essential information to the online platformproviders of online marketplaces, including for purposes of promoting messages on or offering products. That requirement should also be applicable to traders that promote messages on products or services on behalf of brands, based on underlying agreements. Those online platforms should store all information in a secure manner for a reasonable period of time that does not exceed what is necessary, so that it can be accessed, in accordance with the applicable law, including on the protection of personal data, by public authorities and private parties with a legitimate interest, including through the orders to provide information referred to in this Regulation.
2021/07/08
Committee: IMCO
Amendment 439 #
Proposal for a regulation
Article 29 – paragraph 2
2. Where several options are available pursuant to paragraph 1, very large online platforms shall provide an easily and clearly accessible functionality on their online interface allowing the recipient of the service to select and to modify at any time their preferred option for each of the recommender systems that determines the relative order of information presented to them.
2021/07/23
Committee: CULT
Amendment 441 #
Proposal for a regulation
Recital 50
(50) To ensure an efficient and adequate application of that obligation, without imposing any disproportionate burdens, the online platformproviders of online marketplaces covered should make reasonable efforts to verify the reliability of the information provided by the traders concerned, in particular by using freely available official online databases and online interfaces, such as national trade registers and the VAT Information Exchange System45 , or by requesting the traders concerned to provide trustworthy supporting documents, such as copies of identity documents, certified bank statements, company certificates and trade register certificates. They may also use other sources, available for use at a distance, which offer a similar degree of reliability for the purpose of complying with this obligation. However, the online platformproviders of online marketplaces covered should not be required to engage in excessive or costly online fact-finding exercises or to carry out verifications on the spot. Nor should such online platformproviders, which have made the reasonable efforts required by this Regulation, be understood as guaranteeing the reliability of the information towards consumer or other interested parties. Such online platformProviders of online marketplaces should also design and organise their online interface in a user- friendly way that enables traders to comply with their obligations under Union law, in particular the requirements set out in Articles 6 and 8 of Directive 2011/83/EU of the European Parliament and of the Council46 , Article 7 of Directive 2005/29/EC of the European Parliament and of the Council47 and Article 3 of Directive 98/6/EC of the European Parliament and of the Council48 . The online interface should allow traders to provide the information referred to in Article 22a of this Regulation, the information referred to in Article 6 of Directive 2011/83/EU on Consumers Rights, information on sustainability of products, and information allowing for the unequivocal identification of the product or the service, including labelling requirements, in compliance with legislation on product safety and product compliance. __________________ 45 https://ec.europa.eu/taxation_customs/vies/ vieshome.do?selectedLanguage=en 46Directive 2011/83/EU of the European Parliament and of the Council of 25 October 2011 on consumer rights, amending Council Directive 93/13/EEC and Directive 1999/44/EC of the European Parliament and of the Council and repealing Council Directive 85/577/EEC and Directive 97/7/EC of the European Parliament and of the Council 47Directive 2005/29/EC of the European Parliament and of the Council of 11 May 2005 concerning unfair business-to- consumer commercial practices in the internal market and amending Council Directive 84/450/EEC, Directives 97/7/EC, 98/27/EC and 2002/65/EC of the European Parliament and of the Council and Regulation (EC) No 2006/2004 of the European Parliament and of the Council (‘Unfair Commercial Practices Directive’) 48Directive 98/6/EC of the European Parliament and of the Council of 16 February 1998 on consumer protection in the indication of the prices of products offered to consumers
2021/07/08
Committee: IMCO
Amendment 441 #
Proposal for a regulation
Article 29 – paragraph 2 a (new)
2a. Very large online platforms shall ensure that their online interface is designed in such a way that there is no danger of it misleading or manipulating the recipients of the service.
2021/07/23
Committee: CULT
Amendment 449 #
Proposal for a regulation
Recital 50 a (new)
(50a) Providers of online marketplaces should demonstrate their best efforts to prevent the dissemination by traders of illegal products and services. In compliance with the no general monitoring principle, providers should inform recipients when the service or product they have acquired through their services is illegal. Once notified of an illegal product or service as foreseen in Article 14, providers of online marketplaces should take effective and proportionate measures to prevent such products or services from reappearing on their online marketplace.
2021/07/08
Committee: IMCO
Amendment 457 #
Proposal for a regulation
Recital 52
(52) Online advertisement plays an important role in the online environment, including in relation to the provision of the services of online platforms. However, online advertisement can contribute to significant risks, ranging from advertisement that is itself illegal content, to contributing to financial incentives for the publication or amplification of illegal or otherwise harmful content and activities online, or the discriminatory display of advertising with an impact on the equal treatment and opportunities of citizens. In addition to the requirements resulting from Article 6 of Directive 2000/31/EC, online platforms should therefore be required to ensure that the recipients of the service have certain individualised information necessary for them to understand when and on whose behalf the advertisement is displayed. In addition, recipients of the service should have an easy access to information on the main parameters used for determining that specific advertising is to be displayed to them, providing meaningful explanations of the logic used to that end, including when this is based on profiling. The requirements of this Regulation on the provision of information relating to advertisement is without prejudice to the application of the relevant provisions of Regulation (EU) 2016/679, in particular those regarding the right to object, automated individual decision- making, including profiling and specifically the need to obtain consent of the data subject prior to the processing of personal data for targeted advertising. Similarly, it is without prejudice to the provisions laid down in Directive 2002/58/EC in particular those regarding the storage of information in terminal equipment and the access to information stored therein.
2021/07/08
Committee: IMCO
Amendment 467 #
Proposal for a regulation
Recital 53
(53) Given the importance of very large online platforms, due to their reach, in particular as expressed in number of recipients of the service, in facilitating public debate, economic transactions and the dissemination of information, opinions and ideas and in influencing how recipients obtain and communicate information online, it is necessary to impose specific obligations on those platforms, in addition to the obligations applicable to all online platforms. Those additional obligations on very large online platforms are necessary to address those public policy concerns, including regarding misleading information or any other types of harmful content there being no alternative and less restrictive measures that would effectively achieve the same result.
2021/07/08
Committee: IMCO
Amendment 476 #
Proposal for a regulation
Recital 57
(57) Three categories of systemic risks should be assessed in-depth. A first category concerns the risks associated with the misuse of their service through the dissemination of illegal content, such as the dissemination of child sexual abuse material or illegal hate speech, and the conduct of illegal activities, such as the sale of products or services prohibited by Union or national law, including counterfeit products or the display of copyright-infringing content. For example, and without prejudice to the personal responsibility of the recipient of the service of very large online platforms for possible illegality of his or her activity under the applicable law, such dissemination or activities may constitute a significant systematic risk where access to such content may be amplified through accounts with a particularly wide reach. A second category concerns the impact of the service on the exercise of fundamental rights, as protected by the Charter of Fundamental Rights, including the freedom of expression and information, the right to private life, the right to non-discrimination and the rights of the child. Such risks may arise, for example, in relation to the design of the algorithmic systems used by the very large online platform or the misuse of their service through the submission of abusive notices or other methods for silencing speech or hampering competition, hampering competition or the way platforms' terms and conditions including content moderation policies, are enforced, including through automatic means. With respect to this category of risks, a particular attention should be paid to the detrimental effect of intimidation of independent press and the harassment of journalists, in particular women who are more often victims of hateful speech and online threats. These should be considered systemic risk as referred to in Article 26 as they pose threat to democratic values, media freedom, freedom of expression and information, and should be subject to dedicated mitigating measures as referred to in Article 27, and priority notice through trusted flaggers as referred to in Article 19. A third category of risks concerns the intentional and, oftentimes, coordinated manipulation of the platform’s service, with a foreseeable impact on health, fundamental rights, civic discourse, electoral processes, public security and protection of minors, having regard to the need to safeguard public order, protect privacy and fight fraudulent and deceptive commercial practices. Such risks may arise, for example, through the creation of fake accounts, the use of bots, and other automated or partially automated behaviours, which may lead to the rapid and widespread dissemination of information that is illegal content or incompatible with an online platform’s terms and conditions.
2021/07/08
Committee: IMCO
Amendment 484 #
Proposal for a regulation
Recital 58
(58) Very large online platforms should deploy the necessary means to diligently mitigate the systemic risks identified in the risk assessment. Very large online platforms should under such mitigating measures consider, for example, enhancing or otherwise adapting the design and functioning of their content moderation, algorithmic recommender systems and online interfaces, so that they discourage and limit the dissemination of illegal content, adapting their decision-making processes, or adapting their terms and condition and intentional manipulation and exploitation of the service, including amplification of harmful content, adapting their decision-making processes, or adapting their terms and conditions, as well as making content moderation policies and the way they are enforced fully transparent for the users. They may also include corrective measures, such as discontinuing advertising revenue for specific content, or other actions, such as improving the visibility of authoritative information sources. Very large online platforms may reinforce their internal processes or supervision of any of their activities, in particular as regards the detection of systemic risks. They may also initiate or increase cooperation with trusted flaggers, organise training sessions and exchanges with trusted flagger organisations, and cooperate with other service providers, including by initiating or joining existing codes of conduct or other self-regulatory measures. Any measures adopted should respect the due diligence requirements of this Regulation and be effective and appropriate for mitigating the specific risks identified, in the interest of safeguarding public order, protecting privacy and fighting fraudulent and deceptive commercial practices, and should be proportionate in light of the very large online platform’s economic capacity and the need to avoid unnecessary restrictions on the use of their service, taking due account of potential negative effects on the fundamental rights of the recipients of the service.
2021/07/08
Committee: IMCO
Amendment 495 #
Proposal for a regulation
Recital 62
(62) A core part of a very large online platform’s business is the manner in which information is prioritised and presented on its online interface to facilitate and optimise access to information for the recipients of the service. This is done, for example, by algorithmically suggesting, ranking and prioritising information, distinguishing through text or other visual representations, or otherwise curating information provided by recipients. Such recommender systems can have a significant impact on the ability of recipients to retrieve and interact with information online. They also play an important role in the amplification of certain messages, the viral dissemination of information and the stimulation of online behaviour. Moreover, these recommender systems can also impact media consumption and cultural practices of users, and may risk locking them into a bubble without providing them with the possibility to open up to other content. Consequently, very large online platforms should ensure that recipients are appropriately informed, and can influence the information presented to them. They should clearly present the main parameters for such recommender systems in an easily comprehensible manner to ensure that the recipients understand how information is prioritised for them. They should also ensure that the recipients enjoy alternative options for the main parameters, including options that are not based on profiling of the recipient.
2021/07/08
Committee: IMCO
Amendment 501 #
Proposal for a regulation
Recital 63 a (new)
(63a) The practice of very large online platforms to associate advertisement with content uploaded by users, could indirectly lead to the promotion of illegal content, or content that is in breach of their terms and conditions and could risk to considerably damage the brand image of the buyers of advertising space. In order to prevent such practice, the very large online platforms should ensure, including through standard contractual guarantees to the buyers of advertising space, that the content to which they associate advertisements is legal, and compliant with their terms and conditions. Furthermore, the very large online platforms should allow advertisers to have access to the results of audits carried out independently and evaluating the commitments and tools of platforms for protecting the brand image of the buyers of advertising space ("brand safety").
2021/07/08
Committee: IMCO
Amendment 503 #
Proposal for a regulation
Recital 64
(64) In order to appropriately supervise the compliance of very large online platforms with the obligations laid down by this Regulation, the Digital Services Coordinator of establishment or the Commission may require access to or reporting of specific data. Such a requirement may include, for example, the data necessary to assess the risks and possible harms, such as the dissemination of illegal and amplification of harmful content brought about by the platform’s systems, data on the accuracy, functioning and testing of algorithmic systems for content moderation, recommender systems or advertising systems, or data on processes and outputs of content moderation or of internal complaint- handling systems within the meaning of this Regulation. Investigations by researchers on the evolution and severity of online systemic risks are particularly important for bridging information asymmetries and establishing a resilient system of risk mitigation, informing online platforms, Digital Services Coordinators, other competent authorities, the Commission and the public. This Regulation therefore provides a framework for compelling access to data from very large online platforms to vetted researchers. All requirements for access to data under that framework should be proportionate and appropriately protect the rights and legitimate interests, including trade secrets and other confidential information, of the platform and any other parties concerned, including the recipients of the service.
2021/07/08
Committee: IMCO
Amendment 516 #
Proposal for a regulation
Recital 67
(67) The Commission and the Board should encourage the drawing-up of codes of conduct to contribute to the application of this Regulation, as well as the compliance of online platforms with the provisions of these codes. While the implementation of codes of conduct should be measurable and subject to public oversight, this should not impair the voluntary nature of such codes and the freedom of interested parties to decide whether to participate. In certain circumstances, it is important that very large online platforms cooperate in the drawing-up and adhere to specific codes of conduct. Nothing in this Regulation prevents other service providers from adhering to the same standards of due diligence, adopting best practices and benefitting from the guidance provided by the Commission and the Board, by participating in the same codes of conduct.
2021/07/08
Committee: IMCO
Amendment 522 #
Proposal for a regulation
Recital 68
(68) It is appropriate that this Regulation identify certain areas of consideration for such codes of conduct. In particular, risk mitigation measures concerning specific types of illegal content should be explored via self- and co-regulatory agreements. Another area for consideration is the possible negative impacts of systemic risks on society and democracy, such as disinformation, harmful content or manipulative and abusive activities. This includes coordinated operations aimed at amplifying information, including disinformation, such as the use of bots or fake accounts for the creation of fake or misleading information, sometimes with a purpose of obtaining economic gain, which are particularly harmful for vulnerable recipients of the service, such as children. In relation to such areas, adherence to and compliance with a given code of conduct by a very large online platform may be considered as an appropriate risk mitigating measure. The refusal without proper explanations by an online platform of the Commission’s invitation to participate in the application of such a code of conduct could be taken into account, where relevant, when determining whether the online platform has infringed the obligations laid down by this Regulation.
2021/07/08
Committee: IMCO
Amendment 527 #
Proposal for a regulation
Recital 69
(69) The rules on codes of conduct under this Regulation could serve as a basis for already established self-regulatory efforts at Union level, including the Product Safety Pledge, the Memorandum of Understanding against counterfeit goods, the Code of Conduct against illegal hate speech as well as the Code of practice on disinformation. In particular for the latter, since the Commission willhas issued guidance for strengthening the Code of practice on disinformation as announced in the European Democracy Action Plan in May 2021.
2021/07/08
Committee: IMCO
Amendment 632 #
Proposal for a regulation
Article 1 – paragraph 5 – point c
(c) Union law on copyright and related rights, in particular Directive (EU) 2019/790 on Copyright and Related Rights in Digital Single Market;
2021/07/08
Committee: IMCO
Amendment 677 #
Proposal for a regulation
Article 2 – paragraph 1 – point f – indent 3 a (new)
- an ‘online search engine’ as defined in point (5) of Article 2 of Regulation (EU) 2019/1150;
2021/07/08
Committee: IMCO
Amendment 679 #
Proposal for a regulation
Article 2 – paragraph 1 – point f a (new)
(fa) live streaming platform services shall be defined as information society services of which the main or one of the main purposes is to give the public access to audio or video material that is live broadcasted by its users, which it organises and promotes for profit-making purposes;
2021/07/08
Committee: IMCO
Amendment 680 #
Proposal for a regulation
Article 2 – paragraph 1 – point f b (new)
(fb) private messaging services shall be defined as number-independent interpersonal communications services as defined in Article 2(7) of Directive (EU) 2018/1972, excluding transmission of electronic mail as defined in Article 2 (h) of Directive 2002/58/EC;
2021/07/08
Committee: IMCO
Amendment 704 #
Proposal for a regulation
Article 2 – paragraph 1 – point h a (new)
(ha) ‘online marketplace’ means an online platform that allows consumers to conclude distance contracts with other traders or consumers on their platform;
2021/07/08
Committee: IMCO
Amendment 730 #
Proposal for a regulation
Article 2 – paragraph 1 – point p a (new)
(pa) ‘deep fake’ means a generated or manipulated image, audio or video content that appreciably resembles existing persons, objects, places or other entities or events and falsely appears to a person to be authentic or truthful;
2021/07/08
Committee: IMCO
Amendment 772 #
Proposal for a regulation
Article 5 – paragraph 3
3. Paragraph 1 shall not apply with respect to liability under consumer protection law of online platforms allowing consumers to conclude distance contracts with traderproviders of online marketplaces, where such an online platformmarketplace presents the specific item of information or otherwise enables the specific transaction at issue in a way that would lead an average and reasonably well-informed consumer to believe that the information, or the product or service that is the object of the transaction, is provided either by the online platformmarketplace itself or by a recipient of the service who is acting under its authority or control.
2021/07/08
Committee: IMCO
Amendment 785 #
Proposal for a regulation
Article 6 – paragraph 1
Providers of intermediary services shall not be deemed ineligible for the exemptions from liability referred to in Articles 3, 4 and 5 solely because they carry out voluntary own-initiative investigations or other activities aimed at detecting, identifying and removing, or disabling of access to, illegal content, or take the necessary measures to comply with the requirements of Union law, including thoseor national law, in conformity with the Union law, including the EU Charter on Fundamental Rights, and the requirements set out in this Regulation.
2021/07/08
Committee: IMCO
Amendment 803 #
Proposal for a regulation
Article 8 – paragraph 1
1. Providers of intermediary services shall, upon the receipt of an order to act against a specific item of illegal content, issued by the relevant national judicial or administrative authorities, on the basis of the applicable Union or national law, in conformity with Union lawor national law, that is in conformity with Union law, including the EU Charter on Fundamental Rights, inform the authority issuing the order of the effect given to the orders, without undue delay, specifying the action taken and the moment when the action was taken.
2021/07/08
Committee: IMCO
Amendment 835 #
Proposal for a regulation
Article 8 – paragraph 2 – point c
(c) the order is drafted in the language declared by the provider and is sent to the point of contact, appointed by the provider, in accordance with Article 10, or in the official language of the Member State that issues the order against the specific item of illegal content. In such case, the point of contact may request the competent authority to provide translation into the language declared by the provider.
2021/07/08
Committee: IMCO
Amendment 845 #
Proposal for a regulation
Article 8 – paragraph 3
3. The Digital Services Coordinator from the Member State of the judicial or administrative authority issuing the order shall, without undue delay, transmit a copy of the orders referred to in paragraph 1 to all other Digital Services Coordinators through the system established in accordance with Article 67. Where upon receiving the copy of the order, at least three Digital Services Coordinators consider that the order violates Union or national law, that is in conformity with the Union Law, including the Charter, they can object the enforcement of the order to the Board, based on a reasoned statement. Following recommendation of the Board, the Commission may decide whether the order shall be enforced. Where the order to act against a specific item of illegal content under Union or national law has been issued by the national judicial or administrative authority of a Member State that is under an Article 7 procedure for infringement of European values according to Article 2 of TEU, any Digital Service Coordinator may object the order directly to the Commission. The Commission shall assess the objection to the order as a matter of priority and decide whether the order should be enforced as swiftly as possible and no later than 48 hours upon receipt of the objection.
2021/07/08
Committee: IMCO
Amendment 850 #
Proposal for a regulation
Article 8 – paragraph 4
4. The conditions and requirements laid down in this article shall be without prejudice to requirements under national criminal procedural law in conformity with Union law, including the EU Charter on Fundamental Rights.
2021/07/08
Committee: IMCO
Amendment 858 #
Proposal for a regulation
Article 9 – paragraph 1
1. Providers of intermediary services shall, upon receipt of an order to provide a specific item of information about one or more specific individual recipients of the service, issued by the relevant national judicial or administrative authorities on the basis of the applicable Union or national law, in conformity with Union law, inform without undue delay the authority of issuing the order of its receipt and the effect given to the order. Where no effect has been given to the order, providers of intermediary services shall provide without delay the authority of issuing the order with a statement of reasons as to why the order was not given an effect.
2021/07/08
Committee: IMCO
Amendment 872 #
Proposal for a regulation
Article 9 – paragraph 2 – point a – indent 1
— a statement of reasons explaining the objective foraccording to which the information is required and why the requirement to provide the information is necessary and proportionate to determine compliance by the recipients of the intermediary services with applicable Union or national rules, unless such a statement cannot be provided for reasons related to the prevention, investigation, detection and prosecution of criminal offences;
2021/07/08
Committee: IMCO
Amendment 878 #
Proposal for a regulation
Article 9 – paragraph 2 – point c
(c) the order is drafted in the language declared by the provider and is sent to the point of contact appointed by that provider, in accordance with Article 10, or in the official language of the Member State that issues the order against the specific item of illegal content. In such case, the point of contact may request the competent authority to provide translation into the language declared by the provider;
2021/07/08
Committee: IMCO
Amendment 933 #
Proposal for a regulation
Article 12 – paragraph 1
1. Providers of intermediary services shall include information on any restrictions that they impose in relation to the use of their service in respect of information provided by the recipients of the service, in their terms and conditions. That information shall include information on any policies, procedures, measures and tools used for the purpose of content moderation, including algorithmic decision-making and human review. It shall be set out in clear and ensure that their terms and conditions prevent the recipients of their services from providing information that is not compliant with Union law or the law of the Member State where the information is provided. Any additional restrictions that providers of intermediary services may impose in relation to the use of their service and the information provided by the recipients of the service shall be in full compliance with the fundambiguous language and shall be publicly available in an easily accessible formatental rights of the recipients of the services as enshrined in the EU Charter on Fundamental Rights.
2021/07/08
Committee: IMCO
Amendment 935 #
Proposal for a regulation
Article 12 – paragraph 1 a (new)
1a. Providers of intermediary services shall include information on any restrictions that they impose in relation to the use of their service in respect of information provided by the recipients of the service, in their terms and conditions. That information shall include information on any policies, procedures, measures and tools used for the purpose of content moderation, including algorithmic decision-making and human review, and available remedies including applicable alternative dispute resolution mechanisms. It shall be set out in clear and unambiguous language and shall be publicly available in an easily accessible format. Providers of intermediary services shall provide recipients of services with a concise and easily readable summary of the terms and conditions, including information on the available remedies and the possibilities for opt-out, where relevant.
2021/07/08
Committee: IMCO
Amendment 941 #
Proposal for a regulation
Article 12 – paragraph 2
2. Providers of intermediary services shall actpply and enforce the restrictions referred to in paragraph 2 in a diligent, objective and, timely, proportionate manner in applying and enforcing the restrictions referred to in paragraph 1and non- discriminatory manner, with due regard to the rights and legitimate interests of all parties involved, including the applicable fundamental rights of the recipients of the service as enshrined in the Charternational and Union law, including the EU Charter on Fundamental Rights.
2021/07/08
Committee: IMCO
Amendment 976 #
Proposal for a regulation
Article 13 – paragraph 1 – introductory part
1. Providers of intermediary services shall publish, at least once a year, clear, easily accessible, comprehensible, and detailed reports on any content moderation they engaged in during the relevant period. The reports shall be available in searchable archives. Those reports shall include, in particular, information on the following, as applicable:
2021/07/08
Committee: IMCO
Amendment 985 #
Proposal for a regulation
Article 13 – paragraph 1 – point b
(b) the number of notices submitted in accordance with Article 14, categorised by the type of alleged illegal content concerned, the number of notices submitted by trusted flaggers, any action taken pursuant to the notices by differentiating whether the action was taken on the basis of the law or the terms and conditions of the provider, and the average time needed for taking the action;
2021/07/08
Committee: IMCO
Amendment 1000 #
Proposal for a regulation
Article 13 – paragraph 1 a (new)
1a. The information provided shall be broken down per Member State in which services are offered and in the Union as a whole.
2021/07/08
Committee: IMCO
Amendment 1007 #
Proposal for a regulation
Article 13 – paragraph 2
2. Paragraph 1 and 1a shall not apply to providers of intermediary services that qualify as micro or small enterprises within the meaning of the Annex to Recommendation 2003/361/EC.
2021/07/08
Committee: IMCO
Amendment 1020 #
Proposal for a regulation
Chapter III – Section 2 – title
Additional provisions applicable to providers of hosting services, including online platforms and to providers of live streaming platform services and of private messaging services
2021/07/08
Committee: IMCO
Amendment 1022 #
Proposal for a regulation
Article 14 – paragraph 1
1. Providers of hosting services, providers of live streaming platform services and of private messaging services shall put mechanisms in place to allow any individual or entity to notify them of the presence on their service of specific items of information that the individual or entity considers to be illegal content, or content that is in breach with their terms and conditions. Those mechanisms shall be easy to access, user- friendly, and allow for the submission of notices exclusively by electronic means and may include: (a) a clearly identifiable banner or single reporting button, allowing users to notify quickly and easily the providers of these services of illegal content they have encountered; (b) providing information to the users on what is considered illegal content under Union and national law; (c) providing information to the users on available national public tools to signal illegal content to the competent authorities.
2021/07/08
Committee: IMCO
Amendment 1033 #
Proposal for a regulation
Article 14 – paragraph 2 – introductory part
2. The mechanisms referred to in paragraph 1 shall be such as to facilitate the submission of sufficiently precise and adequately substantiated notices, on the basis of which a diligent economic operator can identify the illegality or the breach of the content in question with the terms and conditions. To that end, the providers shall take the necessary measures to enable and facilitate the submission of notices containing all of the following elements:
2021/07/08
Committee: IMCO
Amendment 1038 #
Proposal for a regulation
Article 14 – paragraph 2 – point a
(a) an explanation of the reasons why the individual or entity considers the information in question to be illegal content, or content that is in breach with providers' terms and conditions;
2021/07/08
Committee: IMCO
Amendment 1043 #
Proposal for a regulation
Article 14 – paragraph 2 – point b
(b) a clear indication of the electronic location of that information, in particular the exact URL or URLs, and, where necessary, additional information enabling the identification of the illegal content, or content that is in breach with providers' terms and conditions;
2021/07/08
Committee: IMCO
Amendment 1074 #
Proposal for a regulation
Article 14 – paragraph 6
6. Providers of hosting services, of live streaming platform services and of private messaging services shall process any notices that they receive under the mechanisms referred to in paragraph 1, and take their decisions in respect of the information to which the notices relate, or in respect of the recipient of the service who provided this information, in a timely, diligent non-discriminatory and objective manner. Where they use automated means for that processing or decision-making, they shall include information on such use in the notification referred to in paragraph 4.
2021/07/08
Committee: IMCO
Amendment 1080 #
Proposal for a regulation
Article 14 – paragraph 6 a (new)
6a. Providers of hosting services, of live streaming platform services and of private messaging services shall demonstrate their best efforts to prevent from reappearing content which is identical to another piece of content that has already been identified and removed by them as illegal. The application of this requirement shall not lead to any general monitoring obligation.
2021/07/08
Committee: IMCO
Amendment 1094 #
Proposal for a regulation
Article 15 – paragraph 1
1. Where a provider of hosting services decides to remove or, disable access to or otherwise restrict the visibility of specific items of information provided by the recipients of the service or to suspend or terminate monetary payments related to those items, irrespective of the means used for detecting, identifying or, removing or disabling access to or reducing the visibility of that information and of the reason for its decision, it shall inform the recipient, at the latest at the time of the removal or disabling of access or the restriction of visibility or the suspension or termination of monetization, of the decision and provide a clear and specific statement of reasons for that decision.
2021/07/08
Committee: IMCO
Amendment 1099 #
Proposal for a regulation
Article 15 – paragraph 1 a (new)
1a. When the removing or disabling access to specific items of information is followed by the transmission of these specific items of information in accordance with Article 15a, the requirement to inform the recipient set out in paragraph 1 may be postponed by a period of six weeks in order to avoid interfere with potential ongoing criminal investigations. The period of six weeks can be renewed only following a motivated decision of the competent authority to which the specific items of information had been transmitted.
2021/07/08
Committee: IMCO
Amendment 1105 #
Proposal for a regulation
Article 15 – paragraph 2 – point a
(a) whether the decision entails either the removal of, or the disabling of access to, the restriction of the visibility of, or the demonetisation of, the information and, where relevant, the territorial scope of the disabling of access or the restriction;
2021/07/08
Committee: IMCO
Amendment 1127 #
Proposal for a regulation
Article 15 a (new)
Article 15a Preservation of content and related data, and mandatory transmission of specific items of information 1. Providers of hosting services shall store the illegal content which has been removed or access to which has been disabled as a result of content moderation, or of an order to act against a specific item of illegal content as referred to in Article 8, as well as any related data removed as a consequence of the removal of such illegal content, which are necessary for administrative or judicial review proceedings, including or out-of- court dispute settlement against a decision to remove or disable access to illegal content and related data. 2. The illegal content and related data, as referred to in paragraph 1, shall be stored for six months from the date of removal or disabling. The illegal content shall, upon request from the competent authority or court, be preserved for a further specified period only if and for as long as necessary for ongoing administrative or judicial review proceedings, as referred to in paragraph 1. 3. Providers of hosting services shall ensure that the illegal content and related data stored pursuant to paragraph 1 are subject to appropriate technical and organisational safeguards. Those technical and organisational safeguards shall ensure that the illegal content and related data stored are accessed and processed only for the purposes referred to in paragraph 1, and ensure a high level of security of the personal data concerned. Providers of hosting services shall review and update those safeguards where necessary. 4. Providers of hosting services shall transmit to the competent authorities of the Member States the illegal content which has been removed or access to which has been disabled, whether such removing or disabling access is a result of a voluntary content moderation or of a use of the notification and action mechanism referred to in Article 14. This obligation of transmission applies under the following conditions: (a) illegal content referred to in this paragraph means content which is manifestly illegal and is an offense according to Framework Decision2008/913/JHA and Directive 2011/36/EU; and (b) the competent law enforcement authority to which to transmit such illegal content is that of the Member State of the residence or establishment of the person who made the illegal content available, or, failing that, the law enforcement authority of the Member State in which the provider of hosting services is established or has its legal representative; or, failing that, the provider of hosting services shall inform Europol; (c) when the provider of hosting services is a very large online platform in accordance with Section 4 of Chapter III, it must also, when transmitting the illegal content, add an indicating flag for the illegal content which involve a threat to the life or safety of persons. 5. Each Member State shall notify to the Commission the list of its competent law enforcement authorities as referred to in paragraph 4.
2021/07/08
Committee: IMCO
Amendment 1134 #
Proposal for a regulation
Article 15 b (new)
Article 15b Notification of suspicions of criminal offences 1. Where provider of hosting service becomes aware of any information giving rise to a suspicion that a serious criminal offence involving a threat to the life or safety of persons has taken place, is taking place or is likely to take place, it shall remove or disable the content and promptly inform the law enforcement or judicial authorities of the Member State or Member States concerned of its suspicion and provide all relevant information available. 2. Where the provider of hosting service cannot identify with reasonable certainty the Member State concerned, it shall inform the law enforcement authorities of the Member State in which it is established or has its legal representative or inform Europol. 3. For the purpose of this Article, the Member State concerned shall be the Member State where the offence is suspected to have taken place, be taking place and likely to take place, or the Member State where the suspected offender resides or is located, or the Member State where the victim of the suspected offence resides or is located. 4. For the purpose of this Article, Member States shall notify to the Commission the list of its competent law enforcement or judicial authorities.
2021/07/08
Committee: IMCO
Amendment 1143 #
Proposal for a regulation
Article 17 – paragraph 1 – introductory part
1. Online platforms shall provide recipients of the service, as well as individuals or entities that have submitted a notice for a period of at least six months following the decision referred to in this paragraph, the access to an effective internal complaint-handling system, which enables the complaints to be lodged electronically and free of charge, against the followingdecision taken by the online platform not to act after having received a notice, and against the decisions taken by the online platform on the ground that the information provided by the recipients is illegal content under Union or national law, or incompatible with its terms and conditions:
2021/07/08
Committee: IMCO
Amendment 1155 #
Proposal for a regulation
Article 17 – paragraph 1 – point a
(a) decisions to remove or, disable access to or restrict the visibility of the information;
2021/07/08
Committee: IMCO
Amendment 1167 #
Proposal for a regulation
Article 17 – paragraph 1 – point c a (new)
(ca) decisions to restrict the ability to monetise content provided by the recipients;
2021/07/08
Committee: IMCO
Amendment 1170 #
Proposal for a regulation
Article 17 – paragraph 1 – point c b (new)
(cb) decisions of online marketplaces to suspend the provisions of their services to traders;
2021/07/08
Committee: IMCO
Amendment 1173 #
Proposal for a regulation
Article 17 – paragraph 1 a (new)
1a. When the decision to remove or disable access to the information is followed by the transmission of this information in accordance with Article 15a, the period of at least six months as set out in paragraph 1 shall be considered to start from the day on which the recipient was informed in accordance with Article 15(2).
2021/07/08
Committee: IMCO
Amendment 1193 #
Proposal for a regulation
Article 17 – paragraph 5
5. Online platforms shall ensure that recipients of the service are given the possibility, where necessary, to contact a human interlocutor at the time of the submission of the complaint and that the decisions, referred to in paragraph 4, are not solely taken on the basis of automated means.
2021/07/08
Committee: IMCO
Amendment 1212 #
Proposal for a regulation
Article 18 – paragraph 2 – subparagraph 1 – point a
(a) it is impartial and independent of online platforms and recipients of the service provided by the online platforms and is legally distinct from and functionally independent of the government of the Member State or any other public or private body;
2021/07/08
Committee: IMCO
Amendment 1268 #
Proposal for a regulation
Article 19 – paragraph 2 – point a
(a) it has particular expertise and competence for the purposes of detecting, identifying and notifying illegal content, as well as intentional manipulation and exploitation of the service in the sense of Article 26, paragraph 1(c);
2021/07/08
Committee: IMCO
Amendment 1273 #
Proposal for a regulation
Article 19 – paragraph 2 – point b
(b) it represents collective interests and is independent from any online platform, law enforcement, or other government or relevant commercial entity;
2021/07/08
Committee: IMCO
Amendment 1301 #
Proposal for a regulation
Article 19 – paragraph 4 a (new)
4a. Member States may recognise entities, that were awarded the status of trusted flaggers in another Member State as a trusted flagger on their own territory. Upon request by a Member State, trusted flaggers can be awarded the status of European trusted flagger by the Board, in accordance with Article 48, paragraph 2. The Commission shall keep register of European trusted flaggers.
2021/07/08
Committee: IMCO
Amendment 1313 #
Proposal for a regulation
Article 19 – paragraph 7
7. The Commission, after consulting the Board, mayshall issue guidance to assist online platforms and Digital Services Coordinators in the application of paragraphs 2, 4a, 5 and 6.
2021/07/08
Committee: IMCO
Amendment 1320 #
Proposal for a regulation
Article 20 – paragraph 1
1. Online platforms shall suspend, for a reasonable period of time and after having issued a prior warning, the provision of their services to recipients of the service that frequently provide manifestly illegal content, or content that is in breach with their terms and conditions.
2021/07/08
Committee: IMCO
Amendment 1351 #
Proposal for a regulation
Article 21
Notification of suspicions of criminal 1. aware of any information giving rise to a suspicion that a serious criminal offence involving a threat to the life or safety of persons has taken place, is taking place or is likely to take place, it shall promptly inform the law enforcement or judicial authorities of the Member State or Member States concerned of its suspicion and provide all relevant information available. 2. identify with reasonable certainty the Member State concerned, it shall inform the law enforcement authorities of the Member State in which it is established or has its legal representative or inform Europol. For the purpose of this Article, the Member State concerned shall be the Member State where the offence is suspected to have taken place, be taking place and likely to take place, or the Member State where the suspected offender resides or is located, or the Member State where the victim of the suspected offence resides or is located.Article 21 deleted offences Where an online platform becomes Where the online platform cannot
2021/07/08
Committee: IMCO
Amendment 1369 #
Proposal for a regulation
Article 22 – title
Traceability of traders on online marketplaces
2021/07/08
Committee: IMCO
Amendment 1373 #
Proposal for a regulation
Article 22 – paragraph 1 – introductory part
1. Where an online platform allows consumers to conclude distance contracts with traders, itProviders of online marketplaces shall ensure that traders can only use its services to promote messages on or to offer products or services to consumers located in the Union if, prior to the use of itstheir services, the online platformmarketplaces hasve obtained the following information:
2021/07/08
Committee: IMCO
Amendment 1386 #
Proposal for a regulation
Article 22 – paragraph 1 – point c
(c) the bankpayment account details of the trader, where the trader is a natural person;
2021/07/08
Committee: IMCO
Amendment 1388 #
Proposal for a regulation
Article 22 – paragraph 1 – point d
(d) the name, address, telephone number and electronic mail address of the economic operator, within the meaning ofestablished in the Union and carrying out the tasks in accordance with Article 3(13) and Article 4 of Regulation (EU) 2019/1020 of the European Parliament and the Council51 or [Article XX of the General Product Safety Regulation], or any relevant act of Union law; __________________ 51Regulation (EU) 2019/1020 of the European Parliament and of the Council of 20 June 2019 on market surveillance and compliance of products and amending Directive 2004/42/EC and Regulations (EC) No 765/2008 and (EU) No 305/2011 (OJ L 169, 25.6.2019, p. 1).
2021/07/08
Committee: IMCO
Amendment 1399 #
Proposal for a regulation
Article 22 – paragraph 1 a (new)
1a. Providers of online marketplaces shall require traders to provide the information referred to in points (a) and (e) immediately upon initial registration for its services. Traders shall be required to provide any supplementary material relating to the information requirements set out in Article 22(1) within a reasonable period, and prior to the use of the service and offering of products and services to consumer.
2021/07/08
Committee: IMCO
Amendment 1405 #
Proposal for a regulation
Article 22 – paragraph 2
2. The online platformproviders of online marketplaces shall, upon receiving that information, make reasonable and before allowing traders to use their services, make best efforts to assess whether the information referred to in points (a), (d) and (e) of paragraph 1 is reliablaccurate through the use of any freely accessible official online database or online interface made available by an authorized administrator or a Member States or the Union or through direct requests to the trader to provide supporting documents from reliable sources.
2021/07/08
Committee: IMCO
Amendment 1417 #
Proposal for a regulation
Article 22 – paragraph 3 – subparagraph 1
Where the online platformproviders of online marketplaces obtains indications that any item of information referred to in paragraph 1 obtained from the trader concerned is inaccurate or incomplete, that platformonline marketplace shall request the trader to correct the information in so far as necessary to ensure that all information is accurate and complete, without delay or within the time period set by Union and national law.
2021/07/08
Committee: IMCO
Amendment 1418 #
Proposal for a regulation
Article 22 – paragraph 3 – subparagraph 2
Where the trader fails to correct or complete that information, the online platformproviders of online marketplaces shall suspend the provision of its service to the trader in relations to the offering of products or services to consumers located in the Union until the request is fully complied with.
2021/07/08
Committee: IMCO
Amendment 1424 #
Proposal for a regulation
Article 22 – paragraph 3 a (new)
3a. The providers of online marketplaces shall ensure that traders are given the ability to discuss any information viewed as inaccurate or incomplete directly with a trader before any suspension of services. This may take the form of the internal complaint- handling system under Article 17.
2021/07/08
Committee: IMCO
Amendment 1426 #
Proposal for a regulation
Article 22 – paragraph 3 b (new)
3b. If an online marketplaces rejects an application for services or suspends services to a trader, the trader shall have recourse to the systems under Article 17 and Article 43 of this Regulation.
2021/07/08
Committee: IMCO
Amendment 1428 #
Proposal for a regulation
Article 22 – paragraph 3 c (new)
3c. Traders shall be solely liable for the accuracy the information provided and shall inform without delay the online marketplace of any changes to the information provided.
2021/07/08
Committee: IMCO
Amendment 1436 #
Proposal for a regulation
Article 22 – paragraph 4
4. The online platformproviders of online market places shall store the information obtained pursuant to paragraph 1 and 2 in a secure manner for the duration of their contractual relationship with the trader concerned. They shall subsequently delete the information.
2021/07/08
Committee: IMCO
Amendment 1438 #
Proposal for a regulation
Article 22 – paragraph 5
5. Without prejudice to paragraph 2, the platformroviders of online marketplaces shall only disclose the information to third parties where so required in accordance with the applicable law, including the orders referred to in Article 9 and any orders issued by Member States’ competent authorities or the Commission for the performance of their tasks under this Regulation.
2021/07/08
Committee: IMCO
Amendment 1443 #
Proposal for a regulation
Article 22 – paragraph 6
6. The online platformproviders of online marketplaces shall make the information referred to in points (a), (d), (e) and (f) of paragraph 1 available to the recipients of the service, in a clear, easily accessible and comprehensible manner.
2021/07/08
Committee: IMCO
Amendment 1449 #
Proposal for a regulation
Article 22 – paragraph 7
7. The online platform shall design and organise its online interface in a way that enables traders to comply with their obligations regarding pre-contractual information and product safety information under applicable Union law.deleted
2021/07/08
Committee: IMCO
Amendment 1461 #
Proposal for a regulation
Article 22 a (new)
Article 22a Compliance by design 1. Providers of online marketplaces shall design and organise their online interface in a fair and user-friendly way that enables traders to comply with their obligations regarding pre-contractual information and product safety information under applicable Union law. 2. The online interface shall allow traders to provide in particular the information referred to under paragraph 6 of Article 22, the information referred to in Article 6 of Directive 2011/83/EU on Consumers Rights, information allowing for the unequivocal identification of the product or the service, and, where applicable, information on sustainability of products, information on labelling, including CE marking, according to the Union legislation on product safety and compliance. 3. This Article is without prejudice to additional requirements under other Union acts, including the [General Product Safety Regulation] and [Market Surveillance Regulation]
2021/07/08
Committee: IMCO
Amendment 1472 #
Proposal for a regulation
Article 23 – paragraph 1 – point c a (new)
(ca) the number of advertisements that were removed, labelled or disabled by the online platform and justification of the decisions;
2021/07/08
Committee: IMCO
Amendment 1478 #
Proposal for a regulation
Article 23 – paragraph 4
4. The Commission mayshall adopt implementing acts to establish a set of key performance indicators and lay down templates concerning the form, content and other details of reports pursuant to paragraph 1.
2021/07/08
Committee: IMCO
Amendment 1483 #
Proposal for a regulation
Article 24 – title
Online advertising transparency and control
2021/07/08
Committee: IMCO
Amendment 1487 #
Proposal for a regulation
Article 24 – paragraph 1 – point a
(a) that the information displayed is an advertisementon the interface or parts thereof is an online advertisement, including through prominent and harmonised marking;
2021/07/08
Committee: IMCO
Amendment 1493 #
Proposal for a regulation
Article 24 – paragraph 1 – point c
(c) clear, meaningful and uniform information about the main parameters used to determine the recipient to whom the advertisement is displayed. and the logic involved;
2021/07/08
Committee: IMCO
Amendment 1505 #
Proposal for a regulation
Article 24 – paragraph 1 a (new)
The online platform shall design and organise its online interface in such a way that recipients of the service can easily and efficiently exercise their rights under applicable Union law in relation to the processing of their personal data for each specific advertisement displayed to the data subject on the platform, in particular: (a) to withdraw consent or to object to processing; (b) to obtain access to the personal data concerning the data subject; (c) to obtain rectification of inaccurate personal data concerning the data subject; (d) to obtain erasure of personal data without undue delay; Where a recipient exercises any of these rights, the online platform must inform any parties to whom the personal data concerned in points (a)-(d) have been enclosed in accordance with Article 19 of Regulation (EU) 2016/679.
2021/07/08
Committee: IMCO
Amendment 1513 #
Proposal for a regulation
Article 24 – paragraph 1 b (new)
Where a recipient exercises any of the rights referred to points (a), (c) or (d) in paragraph 2, the online platform must immediately cease displaying advertisements using the personal data concerned or using parameters which were set using this data.
2021/07/08
Committee: IMCO
Amendment 1515 #
Proposal for a regulation
Article 24 – paragraph 1 c (new)
Online platforms that display advertising on their online interfaces shall ensure that advertisers: (a) can request and obtain information on where their advertisements have been placed; (b) can request and obtain information on which broker treated their data; (c) can indicate on which specific location their ads cannot be placed. In case of non-compliance with this provision, advertisers shall have the right to judicial redress.
2021/07/08
Committee: IMCO
Amendment 1526 #
Proposal for a regulation
Chapter III – Section 4 – title
4Additional obligations for very large online platforms, live streaming platforms, private messaging providers and search engines to manage systemic risks
2021/07/08
Committee: IMCO
Amendment 1528 #
Proposal for a regulation
Article 25 – title
Very large online platforms, live streaming platforms, private messaging providers and search engines
2021/07/08
Committee: IMCO
Amendment 1531 #
Proposal for a regulation
Article 25 – paragraph 1
1. This Section shall apply to online platform services, live streaming platform services, private messaging services and search engine services which provide their services to a number of average monthly active recipients of the service in the Union equal to or higher than 45 million, calculated in accordance with the methodology set out in the delegated acts referred to in paragraph 3.
2021/07/08
Committee: IMCO
Amendment 1545 #
Proposal for a regulation
Article 26 – paragraph 1 – introductory part
1. Very large online platforms shall identify, analyse and assess, from the date of application referred to in the second subparagraph of Article 25(4), on an ongoing basis and at least once a year thereafter, any significthe probability and severity of anty systemic risks stemming from the design, intrinsic characteristics, functioning and use made of their services in the Union. The risk assessment shall be broken down per Member State in which services are offered and in the Union as a whole. This risk assessment shall be specific to their services and shall include the following systemic risks:
2021/07/08
Committee: IMCO
Amendment 1555 #
Proposal for a regulation
Article 26 – paragraph 1 – point a
(a) the dissemination of illegal content through their serviand content that is in breach of their terms and conditions through their services, including unsafe and non- compliant products and services, in case of online marketplaces;
2021/07/08
Committee: IMCO
Amendment 1560 #
Proposal for a regulation
Article 26 – paragraph 1 – point a a (new)
(aa) the funding of illegal content, including models based on advertisement;
2021/07/08
Committee: IMCO
Amendment 1564 #
Proposal for a regulation
Article 26 – paragraph 1 – point b
(b) any negative effects for the exercise of any of the fundamental rights listed in the EU Charter on Fundamental Rights , in particular on the fundamental rights to respect for private and family life, freedom of expression and information, the prohibition of discrimination and the rights of the child, as enshrined in Articles 7, 11, 21 and 24 of the Charter respectively;
2021/07/08
Committee: IMCO
Amendment 1573 #
Proposal for a regulation
Article 26 – paragraph 1 – point c
(c) intentional manipulation of their service and amplification of content that is in breach of their terms and conditions, including by means of inauthentic use, such as ‘deep fakes’ or automated exploitation of the service, with an actual or foreseeable negative effect on the protection of public health, minors, democratic values, media freedom and freedom of expression of journalists, as well as their ability to verify facts, civic discourse, or actual or foreseeable effects related to electoral processes and public security.
2021/07/08
Committee: IMCO
Amendment 1584 #
Proposal for a regulation
Article 26 – paragraph 2
2. When conducting risk assessments, very large online platforms shall take into account, in particular, how and whether their content moderation systems, recommender systems and systems for selecting and displaying advertisement influence any of the systemic risks referred to in paragraph 1, including the potentially rapid and wide dissemination of illegal content and of information that is in compatible with their terms and conditions, as well as potential infringement of consumer rights by business active on the platform or platform themselves.
2021/07/08
Committee: IMCO
Amendment 1593 #
Proposal for a regulation
Article 26 – paragraph 2 a (new)
2a. When conducting risk assessments, very large online platforms shall involve representatives of the recipients of the service, representatives of groups potentially impacted by their services, independent experts and civil society organisations. Their involvement shall be tailored to the specific systemic risks that the very large online platform aim to assess.
2021/07/08
Committee: IMCO
Amendment 1601 #
Proposal for a regulation
Article 27 – paragraph 1 – introductory part
1. Very large online platforms shall put in place reasonable, proportionate and effective mitigation measureseasures to mitigate the probability and severity of any, tailored to address the specific systemic risks identified pursuant to Article 26. Such measures may include, where applicable:
2021/07/08
Committee: IMCO
Amendment 1609 #
Proposal for a regulation
Article 27 – paragraph 1 – point a
(a) adapting content moderation or recommender systems, their decision- making processes, design, the features or functioning of their services, or their terms and conditions;
2021/07/08
Committee: IMCO
Amendment 1614 #
Proposal for a regulation
Article 27 – paragraph 1 – point b
(b) targeted measures aimed at limiting the display of and targeting of advertisements in association with the service they provide;
2021/07/08
Committee: IMCO
Amendment 1627 #
Proposal for a regulation
Article 27 – paragraph 1 a (new)
1a. The Board shall evaluate the implementation and effectiveness of mitigating measures undertaken by very large online platforms listed in Article 27(1) and where necessary, may issue recommendations.
2021/07/08
Committee: IMCO
Amendment 1630 #
Proposal for a regulation
Article 27 – paragraph 1 b (new)
1b. Where a very large online platform decides not to put in place any of the mitigating measures listed in Article 27(1), it shall provide a written explanation that describes the reasons why those measures were not put in place, to the Board in view of issuing specific recommendations and to independent auditors for the purposes of the audit report. Following the written explanation of the reasons of the very large online platforms not to put in place mitigating measures, and where necessary, the Board shall issue specific recommendations as to the mitigation measures that very large online platforms shall implement. Very large online platforms shall within one month from receiving of these recommendations, implement the recommended measures, or set out any alternative measures they intend to take to address the identified risks. In case of systemic failure of a very large online platform to take effective mitigating measures and in case of repeated non-compliance with the recommendations, the Board may advise the Commission and the Digital Services Coordinators to impose sanctions.
2021/07/08
Committee: IMCO
Amendment 1631 #
Proposal for a regulation
Article 27 – paragraph 2 – introductory part
2. The Board, in cooperation with the Commission, shall publish comprehensive reports, once a year, which. The reports of the Board shall be broken down per Member State in which the systemic risks occur and in the Union as a whole. The reports shall be published in all the official languages of the Member States of the Union. The reports shall include the following:
2021/07/08
Committee: IMCO
Amendment 1637 #
Proposal for a regulation
Article 27 – paragraph 2 – point a
(a) identification and assessment of the most prominent and recurrenteach of the systemic risks reported by very large online platforms or identified through other information sources, in particular those provided in compliance with Article 31 and 33;
2021/07/08
Committee: IMCO
Amendment 1644 #
Proposal for a regulation
Article 27 – paragraph 3
3. The Commission, in cooperation with the Digital Services Coordinators, mayand following public consultations shall issue general guidelines on the application of paragraph 1 in relation to specific risks, in particular to present best practices and recommend possible measures, having due regard to the possible consequences of the measures on fundamental rights enshrined in the Charter of all parties involved. When preparing those guidelines the Commission shall organise public consultations.
2021/07/08
Committee: IMCO
Amendment 1659 #
Proposal for a regulation
Article 28 – paragraph 1 – point a
(a) the obligations set out in Chapter III, in particular the quality of the identification, analysis and assessment of the risks referred to in Article26, and the necessity, proportionality and effectiveness of the risk mitigation measures referred to in Article 27;
2021/07/08
Committee: IMCO
Amendment 1672 #
Proposal for a regulation
Article 28 – paragraph 2 – point c a (new)
(ca) have been certified by the Commission for the performance of this task;
2021/07/08
Committee: IMCO
Amendment 1676 #
Proposal for a regulation
Article 28 – paragraph 3 – point f a (new)
(fa) a description of specific elements that could not be audited, and an explanation of why these could not be audited;
2021/07/08
Committee: IMCO
Amendment 1678 #
Proposal for a regulation
Article 28 – paragraph 3 – point f b (new)
(fb) where the audit opinion could not reach a conclusion for specific elements within the scope of the audit, a statement of reasons for the failure to reach such conclusion.
2021/07/08
Committee: IMCO
Amendment 1692 #
Proposal for a regulation
Article 29 – paragraph 1
1. Very large online platforms that use recommender systems shall set out in their terms and conditions and on a designated web page that can be directly reached from the very large online platforms’ online interface, in a clear, accessible and easily comprehensible manner for the general public, the main parameters used in their recommender systems, the optimisation goals of their recommender systems as well as any options for the recipients of the service to modify or influence those main parameters that they may have made available, including at least one option which is not based on profiling, within the meaning of Article 4 (4) of Regulation (EU) 2016/679.
2021/07/08
Committee: IMCO
Amendment 1701 #
Proposal for a regulation
Article 29 – paragraph 2
2. Where several options are available pursuant to paragraph 1, very large online platforms shall provide clear and easily accessible functionality on their online interface allowing the recipient of the service to select and to modify at any time their preferred option for each of the recommender systems that determines the relative order of information presented to them.
2021/07/08
Committee: IMCO
Amendment 1704 #
Proposal for a regulation
Article 29 – paragraph 2 a (new)
2a. Online platforms shall ensure that their online interface is designed in such a way that it does not risk misleading or manipulating the recipients of the service.
2021/07/08
Committee: IMCO
Amendment 1709 #
Proposal for a regulation
Article 30 – title
Additional online advertising transparencytransparency for online advertisements and ‘deep fakes’ audiovisual media
2021/07/08
Committee: IMCO
Amendment 1716 #
Proposal for a regulation
Article 30 – paragraph 1
1. Very large online platforms that display advertising on their online interfaces shall compile and make publicly available to relevant authorities and vetted researchers, meeting the requirements of Article 31(4), through application programming interfaces a repository containing the information referred to in paragraph 2, until one year after the advertisement was displayed for the last time on their online interfaces. They shall ensure that the repository does not contain any personal data of the recipients of the service to whom the advertisement was or could have been displayed.
2021/07/08
Committee: IMCO
Amendment 1721 #
Proposal for a regulation
Article 30 – paragraph 2 – point a
(a) the content of the advertisement, including the name of the product, service or brand and the object of the advertisement;
2021/07/08
Committee: IMCO
Amendment 1724 #
Proposal for a regulation
Article 30 – paragraph 2 – point b a (new)
(ba) the natural or legal person who paid for the advertisement;
2021/07/08
Committee: IMCO
Amendment 1732 #
Proposal for a regulation
Article 30 – paragraph 2 – point e
(e) the total number of recipients of the service reached in each country and, where applicable, aggregate numbers for the group or groups of recipients to whom the advertisement was targeted specifically.
2021/07/08
Committee: IMCO
Amendment 1739 #
Proposal for a regulation
Article 30 – paragraph 2 a (new)
2a. The Board shall, after consulting trusted flaggers and vetted researchers, publish guidelines on the structure and organisation on repositories created pursuant to paragraph 1.
2021/07/08
Committee: IMCO
Amendment 1744 #
Proposal for a regulation
Article 30 – paragraph 2 b (new)
2b. Very large online platforms shall make their best effort to detect inauthentic videos (‘deep fakes’). When detecting such videos, they should label them as inauthentic in a way that is clearly visible for the internet user.
2021/07/08
Committee: IMCO
Amendment 1746 #
Proposal for a regulation
Article 30 – paragraph 2 c (new)
2c. Very large online platforms selling advertising for display on their online interface, shall ensure via standard contractual clauses with the purchasers of advertising space that the content with which the advertisement is associated is compliant with the terms and conditions of the platform, or with the law of the Member States where the recipients of the service to whom the advertisement will be displayed is located.
2021/07/08
Committee: IMCO
Amendment 1747 #
Proposal for a regulation
Article 30 – paragraph 2 d (new)
2d. Very large online platforms that display advertising on their online interfaces shall conduct at their own expense, and upon request of advertisers , independent audits performed by organisations complying with the criteria set out in Article 28(2). Such audits shall be based on fair and proportionate conditions agreed between platforms and advertisers, shall be conducted with a reasonable frequency and shall entail: (a) conducting quantitative and qualitative assessment of cases where advertising is associated with illegal content or with content incompatible with platforms’ terms and conditions; (b) monitoring for and detecting of fraudulent use of their services to fund illegal activities; (c) assessing the performance of their tools in terms of brand safety. The audit report shall include opinion on the performance of platforms’ tools in terms of brand safety. Where the audit opinion is not positive, the report shall make operational recommendations to the platforms on specific measures in order to achieve compliance. The platforms shall make available to advertisers, upon request, the results of such audit.
2021/07/08
Committee: IMCO
Amendment 1750 #
Proposal for a regulation
Article 31 – paragraph 1
1. Very large online platforms shall provide the Digital Services Coordinator of establishment or the Commission, upon their reasoned request and within a reasonable periodout delay, specified in the request, full access to data that are necessary to monitor and assess compliance with this Regulation. That Digital Services Coordinator and the Commission shall only use that data for those purposes. With regard to moderation and recommender systems, very large online platforms shall provide upon request the Digital Services Coordinator or the Commission with access to algorithms and associated data that allow the detection of possible biases which could lead to the dissemination of illegal content, or content that is in breach with their terms and conditions, or presents threats to fundamental rights including freedom of expression. Where a bias is detected, very large online platforms shall expeditiously correct it following the recommendations of the Digital Services Coordinator or the Commission. Very large online platforms should be able to demonstrate their compliance at every step of the process pursuant to this Article.
2021/07/08
Committee: IMCO
Amendment 1757 #
Proposal for a regulation
Article 31 – paragraph 2
2. Upon a reasoned request from the Digital Services Coordinator of establishment, three Digital Services Coordinators of destination or the Commission, very large online platforms shall, within a reasonable period, as specified in the request, provide access to data to vetted researchers who meet the requirements in paragraphs 4 of this Article, for the sole purpose of conducting research that contributes to the identification and understanding and mitigation of systemic risks as set out in Articles 26(1) and 27.
2021/07/08
Committee: IMCO
Amendment 1765 #
Proposal for a regulation
Article 31 – paragraph 4
4. In order to be vetted, scientific researchers shall be affiliated with academic institutions, be independent from commercial interests and the very large online platform it seeks data from, have proven records of expertise in the fields related to the risks investigated or related research methodologies, and shall commit and be in a capacity to preserve the specific data security and confidentiality requirements corresponding to each request.
2021/07/08
Committee: IMCO
Amendment 1775 #
Proposal for a regulation
Article 31 – paragraph 6 – introductory part
6. Within 15 days following receipt of a request as referred to in paragraph 1 and 2, a very large online platform may request the Digital Services Coordinator of establishment or the Commission, as applicable, to amend the request, where it considers that it is unable to give access to the data requested because one of following two reasons: for the following reasons: (a) in case of request under paragraph 1, a very large online platform does not have and cannot obtain with reasonable effort access to the data; (b) in case of request under paragraph 2, a very large online platform does not have access to the data or providing access to the data will lead to significant vulnerabilities for the security of its service or the protection of confidential information, in particular trade secrets.
2021/07/08
Committee: IMCO
Amendment 1788 #
Proposal for a regulation
Article 31 – paragraph 7 a (new)
7a. Upon completion of the research envisaged in Article 31(2), the vetted researchers shall make their research publicly available, taking into account the rights and interests of the recipients of the service concerned in compliance with Regulation (EU) 2016/679.
2021/07/08
Committee: IMCO
Amendment 1789 #
Proposal for a regulation
Article 31 – paragraph 7 b (new)
7b. Digital Service Coordinators and the Commission shall, once a year, report the following information: (a) the number of requests made to them as referred to in paragraphs 1 and 2; (b) the number of such requests that have been declined by the Digital Service Coordinator or the Commission and the reasons for which they have been declined; (c) the number of such requests that have been declined by the Digital Service Coordinator or the Commission, including the reasons for which they have been declined, following a request to the Digital Service Coordinator or the Commission from a very large online platform to amend a request as referred to in paragraphs 1 and 2.
2021/07/08
Committee: IMCO
Amendment 1802 #
Proposal for a regulation
Article 33 – paragraph 2 a (new)
2a. The reports shall include content moderation broken down per Member State in which the services are offered and in the Union as a whole and shall be published in the official languages of the Member States of the Union.
2021/07/08
Committee: IMCO
Amendment 1848 #
Proposal for a regulation
Article 35 – paragraph 1
1. The Commission and the Board shall encourage and facilitate the drawing up of codes of conduct at Union level to contribute to the proper application of this Regulation, taking into account in particular the specific challenges of tackling different types of illegal content as defined in Union and national law and systemic risks, in accordance with Union law, in particular on competition and the protection of personal data.
2021/07/08
Committee: IMCO
Amendment 1858 #
Proposal for a regulation
Article 35 – paragraph 2
2. Where significant systemic risk within the meaning of Article 26(1) emerge and concern several very large online platforms, the Commission mayshall invite the very large online platforms concerned, other very large online platforms, other online platforms and other providers of intermediary services, as appropriate, as well as civil society organisations and other interested parties, to participate in the drawing up of codes of conduct, including by setting out commitments to take specific risk mitigation measures, as well as a regular reporting framework on any measures taken and their outcomes.
2021/07/08
Committee: IMCO
Amendment 1867 #
Proposal for a regulation
Article 35 – paragraph 3
3. When giving effect to paragraphs 1 and 2, the Commission and the Board shall aim to ensure that the codes of conduct clearly set out their objectives, contain a set of harmonised key performance indicators to measure the achievement of those objectives and take due account of the needs and interests of all interested parties, including citizens, at Union level. The Commission and the Board shall also aim to ensure that participants report regularly to the Commission and their respective Digital Service Coordinators of establishment on any measures taken and their outcomes, as measured against the key performance indicators that they contain in order to facilitate effective cross-platform monitoring.
2021/07/08
Committee: IMCO
Amendment 1870 #
Proposal for a regulation
Article 35 – paragraph 4
4. The Commission and the Board shall assess whether the codes of conduct meet the aims specified in paragraphs 1 and 3, and shall regularly monitor and evaluate the achievement of their objectives. They shall publish their conclusion, and publish their conclusions. Furthermore, they shall ensure that there is common alert mechanism managed at Unions level to allow for real-time and coordinated responses.
2021/07/08
Committee: IMCO
Amendment 1873 #
Proposal for a regulation
Article 35 – paragraph 5
5. The Board shall regularly monitor and evaluate the achievement of the objectives of the codes of conduct, having regard to the key performance indicators that they may contain. In case of systematic and repetitive failure to comply with the Codes of Conduct, the Board shall as a measure of last resort take a decision to temporary suspend or definitely exclude platforms that do not meet their commitments as a signatory to the Codes of Conduct.
2021/07/08
Committee: IMCO
Amendment 1881 #
Proposal for a regulation
Article 36 – paragraph 1
1. The Commission shall encourage and facilitate the drawing up of codes of conduct at Union level between, online platforms and other relevant service providers, such as providers of online advertising intermediary services or organisations representing recipients of the service and civil society organisations or relevant authorities to contribute to further transparency infor all players in the online advertising value chain. beyond the requirements of Articles 24 and 30.
2021/07/08
Committee: IMCO
Amendment 1888 #
Proposal for a regulation
Article 36 – paragraph 2 – point b a (new)
(ba) the setting-up of unique identifier that will enable advertisers and publishers to identify and track a campaign throughout its lifecycle.
2021/07/08
Committee: IMCO
Amendment 1890 #
Proposal for a regulation
Article 36 – paragraph 3
3. The Commission shall encourage the development of the codes of conduct within one year following the date of application of this Regulation and their application no later than six months after that date. The Commission shall evaluate the application of those codes two years after the application of this Regulation.
2021/07/08
Committee: IMCO
Amendment 1891 #
Proposal for a regulation
Article 36 – paragraph 3 a (new)
3a. The Commission shall encourage all the players in the online advertising value chain to endorse and comply with the commitments stated in the codes of conduct.
2021/07/08
Committee: IMCO
Amendment 1917 #
Proposal for a regulation
Article 38 – paragraph 4 a (new)
4a. Member States shall ensure that the competent authorities have adequate financial and human resources, as well as legal and technical expertise to fulfil their tasks under this Regulation.
2021/07/08
Committee: IMCO
Amendment 1928 #
Proposal for a regulation
Article 40 – paragraph 1 a (new)
1a. By means of derogation from paragraph 1, the Member State in which the consumers have their residence shall have jurisdiction for the purposes of Articles 22, 22a and 22b of this Regulation and the Member State in which the authority issuing the order is situated shall have jurisdiction for the purposes of Articles 8 and 9 of this Regulation.
2021/07/08
Committee: IMCO
Amendment 1939 #
Proposal for a regulation
Article 40 – paragraph 4
4. Paragraphs 1,1a, 2 and 3 are without prejudice to Article 43(2), the second subparagraph of Article 50(4) and the second subparagraph of Article 51(2) and the tasks and powers of the Commission under Section 3.
2021/07/08
Committee: IMCO
Amendment 1954 #
Proposal for a regulation
Article 41 – paragraph 3 a (new)
3a. Following request to the Commission and in cases of infringements that persist, could cause serious harm to recipients of the service, or could seriously affect their fundamental rights, the Digital Services Coordinator of the country of destination may be entitled to additional powers in the framework of joint investigations as referred to in Article 46.
2021/07/08
Committee: IMCO
Amendment 1972 #
Proposal for a regulation
Article 43 – paragraph 1 a (new)
Where the complaint concerns an alleged harm upon the recipients of the service, the Member State where the recipient resides shall have jurisdiction for the purposes of the complaint.
2021/07/08
Committee: IMCO
Amendment 1974 #
Proposal for a regulation
Article 43 a (new)
Article 43a Rights to effective judicial remedies 1. Without prejudice to any available administrative or non-judicial remedy, any recipient of the service or representative organisations shall have the right to an effective judicial remedy where he or she suffered harm as a result of an infringement of Articles 26(1) and 27(1). 2. In determining whether the very large online platform has complied with its obligations under Article 27(1), and in light of the principle of proportionality, the availability of suitable and effective measures shall be taken into account. 3. Such proceedings may be brought before the courts of the Member State where the recipient of the service has his or her habitual residence. 4. Without prejudice to any other administrative or non-judicial remedy, any recipients of the service or representative organisations shall have the right to an effective judicial remedy where the Digital Service Coordinator which is competent pursuant to Articles 40 and 43 does not handle a complaint or does not inform the recipient of the service within three months on the progress or outcome of the complaint lodged pursuant to Article 43. Proceedings against a Digital Services Coordinator under paragraph 4 shall be brought before the courts of the Member State where the Digital Services Coordinator is established.
2021/07/08
Committee: IMCO
Amendment 1986 #
Proposal for a regulation
Article 45 – paragraph 1 – subparagraph 2
Where the Board has reasons to suspect that a provider of intermediary services infringed this Regulation in a manner involving at least three Member States, it may recommendshall request the Digital Services Coordinator of establishment to assess the matter and take the necessary investigatory and enforcement measures to ensure compliance with this Regulation.
2021/07/08
Committee: IMCO
Amendment 1989 #
Proposal for a regulation
Article 45 – paragraph 2 – introductory part
2. A request or recommendation pursuant to paragraph 1 shall at least indicate:
2021/07/08
Committee: IMCO
Amendment 1996 #
Proposal for a regulation
Article 45 – paragraph 3
3. The Digital Services Coordinator of establishment shall take into utmost account the request or recommendation pursuant to paragraph 1 and assess the matter in view of taking specific investigatory or enforcement measures to ensure compliance without undue delay. Where it considers that it has insufficient information to act upon the request or recommendation and has reasons to consider that the Digital Services Coordinator that sent the request, and tor the Board, could provide additional information, it may request such information. The time period laid down in paragraph 4 shall be suspended until that additional information is provided.
2021/07/08
Committee: IMCO
Amendment 1998 #
Proposal for a regulation
Article 45 – paragraph 4
4. The Digital Services Coordinator of establishment shall, without undue delay and in any event not later than two months following receipt of the request or recommendation, communicate to the Digital Services Coordinator that sent the request, or the Board, its assessment of the suspected infringement, or that of any other competent authority pursuant to national law where relevant, and an explanation of any investigatory or enforcement measures taken or envisaged in relation thereto and a statement of reason in case of decision, following its investigation, not to take measures to ensure compliance with this Regulation.
2021/07/08
Committee: IMCO
Amendment 2002 #
Proposal for a regulation
Article 45 – paragraph 5
5. Where the Digital Services Coordinator that sent the request, or, where appropriate, the Board, did not receive a reply within the time period laid down in paragraph 4 or where it does not agree with the assessment of the Digital Services Coordinator of establishment, it may refer the matter to the Commission and the Digital Services Coordinators, providing all relevant information. That information shall include at least the request or recommendation sent to the Digital Services Coordinator of establishment, any additional information provided pursuant to paragraph 3 and the communication referred to in paragraph 4.
2021/07/08
Committee: IMCO
Amendment 2006 #
Proposal for a regulation
Article 45 – paragraph 6
6. The Commission , in cooperation with the Digital Services Coordinators shall assess the matter within three months following the referral of the matter pursuant to paragraph 5, after having consulted the Digital Services Coordinator of establishment and, unless it referred the matter itself, the Board.
2021/07/08
Committee: IMCO
Amendment 2010 #
Proposal for a regulation
Article 45 – paragraph 7
7. Where, pursuant to paragraph 6, the Commission in cooperation with the Digital Services Coordinators concludes that the assessment or the investigatory or enforcement measures taken or envisaged pursuant to paragraph 4 are incompatible with this Regulation, it shall request the Digital Service Coordinator of establishment to further assess the matter and take the necessary investigatory or enforcement measures to ensure compliance with this Regulation, and to inform it about those measures taken within two months from that request. Where the Digital Services Coordinator of establishment fails to comply with the request to take the necessary measures before the end of the two months period, the Commission shall reallocate the case without delay to the Digital Services Coordinator initiating the request.
2021/07/08
Committee: IMCO
Amendment 2014 #
Proposal for a regulation
Article 46 – title
Joint investigations, cooperation among Digital Services Coordinators and requests for Commission intervention
2021/07/08
Committee: IMCO
Amendment 2020 #
Proposal for a regulation
Article 46 – paragraph 1 a (new)
1a. Where Digital Services Coordinator of the country of destination considers that an alleged infringement exist and causes serious harm to a large number of recipients of the service in that Member States, or could seriously affect their fundamental rights, it may request to the Commission to set up joint investigations between Digital Services Coordinator of country of establishment and the requesting Digital Services Coordinator of country of destination.
2021/07/08
Committee: IMCO
Amendment 2021 #
Proposal for a regulation
Article 46 – paragraph 1 b (new)
1b. The Commission, in cooperation with the Digital Services Coordinators, shall assess such request and following positive opinion of the Board shall set up a joint investigation where the Digital Services Coordinator of the country of destination can be entitled to exercise the following additional powers with respect to the provider of intermediary services concerned by the alleged infringement: (a) to obtain access to the confidential version of the reports published by the intermediary service providers referred to in Article 13 and where applicable in Articles 23 and 24, as well as to the annual reports drawn up by the other competent authorities pursuant to Article 44; (b) to obtain access to data collected by the Digital Services Coordinator of the country of establishment for the purpose of supervision of that provider on the territory of the Digital Services Coordinator of the country of destination; (c) to initiate proceedings and assess the matter in view of taking specific investigatory or enforcement measures to ensure compliance, where the suspected seriousness of the infringement would require immediate response that would not allow for the provisions of Article 45 to apply; and (d) to request interim measures, as referred to in Article 41(2)(e).
2021/07/08
Committee: IMCO
Amendment 2022 #
Proposal for a regulation
Article 46 – paragraph 1 c (new)
1c. The Commission decision setting up the joint investigation shall define a deadline by when Digital Services Coordinator of the country of establishment and Digital Services Coordinator launching the request pursuant to paragraph 2 shall agree on a common position on the joint investigation, and where applicable on the enforcement measures to be adopted. If no agreement is reached within this deadline, the case shall be referred to the Commission pursuant to Article 45(5).
2021/07/08
Committee: IMCO
Amendment 2082 #
Proposal for a regulation
Article 49 – paragraph 1 – point c a (new)
(ca) issue specific recommendations for the implementation of Article 27 and advise on possible application of sanctions in case of repeated non-compliance;
2021/07/08
Committee: IMCO
Amendment 2100 #
Proposal for a regulation
Article 50 – paragraph 1 – subparagraph 2
The Commission acting on its own initiative, or the Board acting on its own initiative or upon request of at least three Digital Services Coordinators of destination, mayshall, where it has reasons to suspect that a very large online platform infringed any of those provisions, recommend the Digital Services Coordinator of establishment to investigate the suspected infringement with a view to that Digital Services Coordinator adopting such a decision within a reasonable time period.
2021/07/08
Committee: IMCO
Amendment 2119 #
Proposal for a regulation
Article 51 – paragraph 1 – introductory part
1. The Commission, acting either upon the Board’s recommendation or on its own initiative after consulting the Board, mayshall initiate proceedings in view of the possible adoption of decisions pursuant to Articles 58 and 59 in respect of the relevant conduct by the very large online platform that:
2021/07/08
Committee: IMCO
Amendment 2131 #
Proposal for a regulation
Article 51 – paragraph 2 – subparagraph 1
Wheren the Commission decides to initiates proceedings pursuant to paragraph 1, it shall notify all Digital Services Coordinators, the Board and the very large online platform concerned.
2021/07/08
Committee: IMCO