REPORT on the protection of minors online

4.11.2025 - (2025/2060(INI))

Committee on the Internal Market and Consumer Protection
Rapporteur: Christel Schaldemose(


Procedure : 2025/2060(INI)
Document stages in plenary
Document selected :  
A10-0213/2025

MOTION FOR A EUROPEAN PARLIAMENT RESOLUTION

on the protection of minors online

(2025/2060(INI))

The European Parliament,

 having regard to Regulation (EU) 2022/2065 of the European Parliament and of the Council of 19 October 2022 on a Single Market For Digital Services and amending Directive 2000/31/EC (Digital Services Act)[1],

 having regard to its resolution of 12 December 2023 on addictive design of online services and consumer protection in the EU single market[2],

 having regard to its resolution of 12 March 2009 on the protection of consumers, in particular minors, in respect of the use of video games[3],

 having regard to its resolution of 18 January 2023 on consumer protection in online video games: a European Single Market approach[4],

 having regard to the UN Convention on the Rights of the Child of 20 November 1989 and to General Comment No. 25 (2021) of the UN Committee on the Rights of the Child of 2 March 2021 on children’s rights in relation to the digital environment,

 having regard to Regulation (EU) 2024/1689 of the European Parliament and of the Council of 13 June 2024 laying down harmonised rules on artificial intelligence and amending Regulations (EC) No 300/2008, (EU) No 167/2013, (EU) No 168/2013, (EU) 2018/858, (EU) 2018/1139 and (EU) 2019/2144 and Directives 2014/90/EU, (EU) 2016/797 and (EU) 2020/1828 (Artificial Intelligence Act)[5],

 having regard to Directive 2005/29/EC of the European Parliament and of the Council of 11 May 2005 concerning unfair business-to-consumer commercial practices in the internal market and amending Council Directive 84/450/EEC, Directives 97/7/EC, 98/27/EC and 2002/65/EC of the European Parliament and of the Council and Regulation (EC) No 2006/2004 of the European Parliament and of the Council (‘Unfair Commercial Practices Directive’)[6],

 having regard to the Commission’s fitness check of EU consumer law on digital fairness[7],

 having regard to the Commission notice of 17 December 2021 entitled ‘Guidance on the interpretation and application of Directive 2005/29/EC of the European Parliament and of the Council concerning unfair business-to-consumer commercial practices in the internal market[8] (the UCPD Guidance),

 having regard to the Commission communication of 7 October 2025 entitled ‘Guidelines on measures to ensure a high level of privacy, safety and security for minors online, pursuant to Article 28(4) of Regulation (EU) 2022/2065’[9],

 having regard to Regulation (EU) 2017/2394 of the European Parliament and of the Council of 12 December 2017 on cooperation between national authorities responsible for the enforcement of consumer protection laws and repealing Regulation (EC) No 2006/2004[10] (Consumer Protection Cooperation Regulation),

 having regard to Directive (EU) 2018/1808 of the European Parliament and of the Council of 14 November 2018 amending Directive 2010/13/EU on the coordination of certain provisions laid down by law, regulation or administrative action in Member States concerning the provision of audiovisual media services (Audiovisual Media Services Directive) in view of changing market realities[11],

 having regard to its resolution of 9 May 2023 on the implementation of the revised Audiovisual Media Services Directive[12],

 having regard to the Commission communication of 11 May 2022 entitled ‘A Digital Decade for children and youth: the new European strategy for a better internet for kids (BIK+)’ (COM(2022)0212),

 having regard to Regulation (EU) 2024/1183 of the European Parliament and of the Council of 11 April 2024 amending Regulation (EU) No 910/2014, as regards establishing the European Digital Identity Framework[13] (eIDAS2 Regulation),

 having regard to the Commission proposal of 6 February 2024 for a directive of the European Parliament and of the Council on combating child sexual abuse and sexual exploitation of children and child sexual abuse material and replacing Council Framework Decision 2004/68/JHA (COM(2024)0060),

 having regard to Regulation (EU) 2016/679 of the European Parliament and of the Council of 27 April 2016 on the protection of natural persons with regard to the processing of personal data and on the free movement of such data, and repealing Directive 95/46/EC (General Data Protection Regulation)[14],

 having regard to Rule 55 of its Rules of Procedure,

 having regard to the opinion of the Committee on Culture and Education,

 having regard to the report of the Committee on the Internal Market and Consumer Protection (A10-0213/2025),

A. whereas 97 % of young people use the internet daily; whereas 78 % of 13 to 17-year-olds report checking their devices at least hourly and 46 % report checking them almost constantly[15]; whereas 16 to 24-year-olds spend an average of over seven hours a day on the internet[16]; whereas more than half of EU consumers regularly play video games and among children aged from 11 to 14, that share is as high as 84 %[17]; whereas among minors that play video games, 64 % of them spend an average of EUR 1-20 per month on games, while there has been an overall average monthly spending increase by children and consumers in general, from EUR 33 in 2020 to EUR 39 in 2023[18];

B. whereas one in four children and young people display ‘problematic’ or ‘dysfunctional’ smartphone use, meaning behavioural patterns mirroring addiction[19]; whereas minors are more vulnerable to AI-generated content, AI companions and chatbots, as their cognitive abilities are still developing; whereas research shows that habitual checking of social media in teenagers may be associated with changes in neural sensitivity as regards anticipation of social rewards and punishments; whereas research suggests that the rise in mental health problems in adolescents may be related to excessive social media use, and social media pressure has been identified as one of the top five causes of mental health difficulties for minors, such as heightened anxiety and depression; whereas the World Health Organization recognises gaming addiction as a mental health disorder; whereas studies have shown that excessive social media can impact brain development and brain capacity of minors; whereas correlations have been found between excessive social media use and lack of impulse control;

C. whereas citizens, especially minors, are increasingly engaging with social media, video- and content-sharing platforms, and online aggregators, including for news consumption; whereas public figures and influencers active on these online platforms have gained considerable influence in shaping public discourse; whereas many users, including minors, find it challenging to distinguish between true and false information in online news; whereas ensuring the responsible operation and design of these platforms, alongside the promotion of critical thinking and media and digital literacy among minors, must be a priority;

D. whereas digital education and media literacy are essential for strengthening the resilience of minors to online risks, including disinformation and harmful content, while equipping them with the skills to navigate, contribute to and shape the digital world responsibly and confidently; whereas comprehensive educational programmes and awareness campaigns are essential to inform minors, parents, guardians and educators about risks and safe practices online; whereas this includes targeted measures for minors, parents, guardians and educators, prevention techniques, awareness-raising and media and digital literacy campaigns, and information on tools to limit access to age-inappropriate or harmful content;

E. whereas digital technologies and online services can offer benefits for minors by opening up new opportunities for performance, creativity and expression and supporting education; whereas it is essential to guarantee minors’ right of access of to the internet as it contributes to their freedom of expression, learning, information and opinion; whereas minors should not be passive technology consumers, but be actively in control of the technologies they use; whereas media and digital literacy are simultaneously key enablers of children’s active, creative and civic participation online, equipping them with the competences to actively take part in the digital world in a responsible and confident manner, beyond mere risk prevention;

F. whereas the influencer advertising market in Europe has witnessed significant growth in recent years, driven by the increasing importance of social media platforms, a rise in digital advertising budgets and the increasing preference of brands for more authentic forms of marketing; whereas as businesses strive to engage with younger, tech-savvy consumers, influencer marketing has emerged as a key strategy for building brand awareness, trust and credibility; whereas the European market for influencer marketing is projected to grow at a compound annual growth rate of around 12 % over the next five years, and by 2030 the market size is expected to surpass EUR 6.4 billion, demonstrating robust growth prospects; whereas only 20 % of influencers systematically indicate the commercial nature of the content they share;

G. whereas more and more minors are becoming content creators or kidfluencers, sometimes involving commercialisation or early public exposure, which can lead to exploitation without consent;

H. whereas at present, no EU legislation comprehensively addresses the protection of minors across the digital ecosystems; whereas minors should enjoy equal levels of protection regardless of where they live; whereas online platforms and applications have different purposes and services with various designs, interfaces and features, which can lead to different risks to minors, and, therefore, proportionality and risk-based assessment should be applied; whereas minors should enjoy a higher level of protection online, which requires a harmonised EU approach; whereas ensuring higher levels of protection of minors needs to be balanced with respecting their right to access information and participate in online communities;

I. whereas the Digital Services Act (DSA) recognises the protection of minors as an important policy objective of the Union and includes key provisions to ensure better protection of minors online, such as Article 14(3), Article 28, Article 34(1)(d), Article 35(1)(j) and Article 44(1)(j); whereas the Commission guidelines on measures to ensure a high level of privacy, safety and security for minors online, pursuant to Article 28(4) of Regulation (EU) 2022/2065, further clarify the rules on the protection of minors online;

J. whereas the Audiovisual Media Services Directive (AVMSD) identifies the protection of minors as a key policy objective and lays down provisions to strengthen their online safety, notably on video-sharing platforms;

K. whereas the Pan-European Game Information (PEGI) system, as a European video game content rating system, can help consumers – especially parents – make informed decisions about the suitability of video games and apps for different age groups;

L. whereas evidence from several countries in the EU shows that minors have access to pornographic content, sometimes involuntarily, at a very young age; whereas the viewing of pornography can expose children to violent and age-inappropriate content, which has been proven to increase aggression and sexual violence and may trigger harmful sexual behaviours among children and teenagers;

M. whereas the DSA recognises that parental control tools are among risk mitigation measures for online platforms accessible to minors, as provided for in Article 35(1)(j), and thus play a role in managing the risks to minors’ safety, privacy and well-being in the digital environment;

N. whereas the number of young women that are subject to gender-based violence is increasing year by year, in particular in the online sphere; whereas the lack of action and the persistence of online abuse pushes young women out of the digital sphere and impacts their freedom of expression;

O. whereas there is an urgent need to effectively implement and enforce current EU legislation and to address outstanding challenges in order to create a safe online environment for all users, in particular minors; whereas collaboration with Member States is necessary to ensure consistent implementation and enforcement of protective measures across the EU; whereas the enforcement of existing EU legislation remains under-resourced across Member States, resulting in a fragmented implementation and disparities regarding online protection needed for minors; whereas international cooperation with non-EU countries and global organisations is essential to address the cross-border nature of online risks and harms;

P. whereas several Member States have called for a common approach on digital majority and have already adopted national laws in this respect;

Q. whereas the protection of minors’ data is of utmost importance, including ensuring that their data is not misused or exploited, and that a high level of minors’ privacy, safety and security is safeguarded in all online interactions, in line with the DSA;

R. whereas the freedom of expression and of information, as guaranteed by the DSA, does not impede the implementation of effective measures for the protection of minors online;

S. whereas the Commission has committed, in its mission letters and hearings, to proposing a digital fairness act that would tackle unethical techniques and commercial practices related to dark patterns, marketing by social media influencers, the addictive design of digital products and services, and online profiling, especially when consumer vulnerabilities are exploited for commercial purposes for all consumers, including minors;

Implementation and enforcement of existing legislation

1. Stresses that the DSA requires online platforms that are accessible to minors to put in place appropriate and proportionate measures to ensure that minors benefit from a high level of privacy, safety and security when using their services;

2. Takes note of the Commission guidelines on measures to ensure a high level of privacy, safety and security for minors online, pursuant to Article 28(4) of the DSA; welcomes the broad scope of the guidelines, which include principles such as privacy and security by design and practical guidance on the application of assurance mechanisms following a risk-based approach, including the recommendation for high-risk platforms to adopt accurate, robust and privacy-preserving age verification mechanisms; welcomes, furthermore, the guidelines’ support for protective default settings, the reaffirmed responsibility of app providers, regardless of delegation to operating system (OS) operators, and the promotion of existing tools such as the online safety-by-design codes and Child Rights Impact Assessments;

3. Considers that the Commission guidelines could be further improved, in particular regarding the protection of minors from addictive design, profiling-based recommendations and engagement-based recommendations; notes, furthermore, that these guidelines are not legally binding and that further legislative action might be required;

4. Stresses the importance of the swift implementation and effective enforcement of the DSA and other relevant legislation, including the AVMSD, and its full implementation at national level, in order to achieve the objective of guaranteeing a safe online environment for minors; welcomes the Commission’s investigations under the DSA in the field of the protection of minors online, which focus, in particular, on addictive behaviour, the ‘rabbit hole’ phenomenon, age assurance mechanisms and recommender systems; urges the Commission to conclude its investigations imminently and to take any additional steps needed, including fines and effective corrective measures, in accordance with the rules established under the DSA;

5. Recalls that the Commission, on 27 May 2025, initiated formal proceedings against major pornographic platforms for breaches of obligations arising from the DSA; emphasises that the Commission’s investigations specifically focus on the protection of minors online and the lack of effective measures to verify users’ age; recalls that pornographic content, as well as violent content and content that perpetuates misogynistic, racist or homophobic views, can cause extremely serious physical, social and emotional consequences for minors, with harmful effects on their psychological development;

6. Stresses the importance of coordinated actions of the Consumer Protection Cooperation (CPC) Network in the field of the protection of minors online, which focus, in particular, on direct appeals to children in advertisements, the use of aggressive commercial practices, the lack of clear and transparent information and influencer marketing;

7. Expresses concern about the spread of dangerous trends and challenges on online platforms, including targeted disinformation campaigns, noting that these could harm minors; highlights the lack of awareness among parents and guardians, who often have limited capacity to react before harm is done; urges online platforms to proactively assess risks for minors, in line with the existing legislation, and detect emerging dangerous trends and challenges; calls on the Commission to establish, in cooperation with the Safer Internet Centres, an efficient rapid alert mechanism at EU level to detect such issues and to require platforms to mitigate these issues; calls on the Commission to propose that platforms display crisis resource panels with contact numbers for third-party support services, such as eating disorder or suicide prevention charities;

8. Expresses concern about the recruitment of minors by criminal networks on online platforms; expresses concern about the lack of more ambitious mitigation measures to protect minors, in particular when it comes to notice and action, hyper-personalised and engagement-based recommendation algorithms leading to addictive behaviour, and dark patterns; notes with alarm the recent trend by some large online platforms indicating a relaxation of strict content moderation practices on their platforms; expresses deep concern that such a move may increase the exposure of minors and other vulnerable users to harmful, illegal or inappropriate content, thereby undermining efforts to ensure a safe online environment;

9. Highlights the importance of risk assessment provisions in the DSA and stresses that the risk assessments carried out by very large online platforms and very large online search engines under the DSA with regard to the safety of minors online are often inadequate, do not go into sufficient detail beyond publicly available information, lack proper independent oversight and underestimate the risks posed to minors; calls on the Commission to make full use of the tools available under the DSA to address these issues and to make sure that obligations regarding risk assessments and related mitigation measures are fully respected; calls on the Commission, as part of its enforcement powers under the DSA, to also verify that online platforms are complying with their own terms and conditions; calls on the Commission to consider, as a last resort, banning persistently non-compliant sites or applications; calls on the Commission to provide guidance, criteria, benchmarks and a clear framework for independent audits;

10. Calls on the Commission to make risk assessments more robust by specifying the information, data and assessments that must be included in systemic risk assessments, such as revenue data from personalised versus non-personalised ads for children and an assessment of any risk to minors and the effect of algorithms on minors’ user experience, in particular how algorithms may expose minors to illegal content or addictive design features; stresses the importance of including minors, guardians and all relevant experts in risk assessments in order to ensure that the best interests and needs of minors are duly taken into consideration when designing the service; expresses major concerns about the ongoing failure of major digital platforms to adequately protect minors on their services, as exposed by whistle-blowers who have revealed serious breaches and harmful practices; calls on the Commission to actively engage with whistle-blowers who have disclosed concerns related to the protection of minors on digital platforms, in order to benefit from their expertise and insights with a view to improving the effectiveness of EU policies and enforcement mechanisms aimed at safeguarding minors online;

11. Recalls that the Commission and the Member States have a shared responsibility when it comes to the enforcement of the DSA; expresses concern that there are significant delays in designating and/or empowering Digital Services Coordinators (DSCs) in several Member States, which significantly undermines the protection of consumers, including minors, across the EU; welcomes the infringement procedures opened against a number of Member States that have failed to designate and/or empower DSCs; urges the Member States that have not yet done so to designate and empower DSCs without delay and to allocate adequate resources to DSCs and other relevant authorities involved in DSA enforcement; stresses the need to ensure, in the multiannual financial framework, that authorities have all the necessary resources to carry out their work properly and effectively;

12. Calls on the Commission to ensure a harmonised approach to enforcement actions by Member States, including through harmonised operational procedures for DSCs, in order to prevent fragmentation in the internal market; invites the Commission, the Board for Digital Services Coordinators and the relevant national authorities to closely monitor and evaluate the effectiveness of measures taken by online platforms to protect minors; urges the Commission to facilitate cooperation between DSCs and CPC authorities at Member State level and provide guidance to help Member States to structure this cooperation;

13. Welcomes the enforcement efforts made by the CPC Network in the field of consumer protection, notably e-commerce and video games, to protect minors from harmful practices, as well as the continued dialogue with the video game sector; urges the Commission to propose a revision of the CPC Regulation to ensure that the Commission has centralised investigative and enforcement powers across the EU in cases of widespread infringement of EU consumer laws;

14. Welcomes the Commission’s ‘digital fairness’ fitness check of consumer law; underlines that some issues remain unaddressed concerning the protection of consumers online, with minors identified as particularly vulnerable to unfair and manipulative online practices; calls on the Commission to strengthen the protection of minors online through the upcoming digital fairness act; believes that persuasive technologies used by online actors, such as targeted advertising, influencer marketing, addictive design, loot boxes, in-app currencies in video games, and dark patterns, should fall under the digital fairness act, which should close legal loopholes and be consistent with current legal instruments in order to better protect consumers online, in particular minors, taking into account the need to avoid unnecessary regulatory burdens;

Age verification

15. Takes note of the fact that there is currently a fragmented approach to age assurance (age verification, age estimation and self-declaration) measures across the EU, and that this leads to fragmentation in the internal market; acknowledges that some Member States have implemented advanced measures to enhance the protection of minors online; regrets that the different national approaches may result in an uneven level of protection of minors online across the EU; acknowledges the importance of the ongoing discussions at both national and EU level on this issue, with a view to determining how to best address this challenge at EU level;

16. Underlines that cultural norms, societal values and public sensitivities regarding digital tools and data use vary significantly across Member States, which influences how societies perceive, accept and trust measures such as age verification; stresses that these differences must be carefully considered to ensure that any European-level initiatives respect national contexts and public trust, while providing effective tools to protect minors;

17. Welcomes the Commission’s risk-based approach taken in its guidelines for effective, proportionate and privacy-preserving age assurance mechanisms to protect minors online, in accordance with the DSA; highlights that the Commission considers the use of access restrictions supported by age verification methods an appropriate and proportionate measure to ensure a high level of privacy, safety and security for minors; calls on the Commission to present, if necessary, appropriate legislative measures to provide legal certainty and ensure a harmonised approach for safe and reliable age assurance mechanisms, taking into account the Commission guidelines on the protection of minors throughout the EU, and guaranteeing an equal level of protection of minors in the Union;

18. Notes, however, the concerns related to age assurance mechanisms regarding possible unintended consequences against the freedoms and rights of the child; calls, therefore, for any legislative measures related to age assurance mechanisms to be thoroughly assessed for their impact on fundamental rights, to avoid enshrining surveillance practices, and to ensure that any solutions proposed prioritise the most effective and privacy-preserving measures and are adopted only where strictly necessary, proportionate and accompanied by robust safeguards;

19. Calls for the Commission to ensure consistency in the area of minors’ safety online across the various pieces of EU legislation in this area;

20. Stresses that age assurance mechanisms are not a silver bullet for protecting minors from the many risks they face online; emphasises that such tools cannot remove platforms’ accountability for ensuring their products are safe and appropriate for minors by design and by default; encourages the development of versions of social media services that are age-appropriate for minors;

21. Highlights the initiative of Commission President Ursula von der Leyen to create a panel of experts to swiftly assess, among other issues, the question of digital age limits and which online services should be covered;

22. Considers that, when requiring confirmation of users’ age, the Commission should only require the use of the most accurate, robust and privacy-preserving mechanisms; considers that such mechanisms should be systemically implemented for online platforms or digital services selling products or services that fall under an age restriction under EU or national legislation and for social media platforms, video-sharing platforms and AI companions, which present risk to minors;

23. Encourages collaboration between the Member States that have already developed digital tools for effective age verification and the Commission in order to develop a strong and effective age verification tool at EU level, ensuring full respect for the technical specifications derived from the eIDAS2 Regulation;

24. Welcomes the Commission’s efforts to provide an EU-wide solution by developing an age-verification app and the EU Digital Identity (eID) Wallet; recommends that the Commission further invest in privacy-preserving systems, such as zero-knowledge proof protocols, building on the principles of minimal exposure and data minimisation; calls on the Commission to ensure that age assurance mechanisms that are deployed are accurate, effective, reliable, robust, non-intrusive, privacy-preserving and non-discriminatory by design, and do not incentivise the development of technologies that rely on the surveillance of users, in particular minors, nor on the deployment of other unethical technological practices;

25. Emphasises that individuals should be granted access to online platforms that use the easiest, safest and most privacy-friendly option by default; considers that, to date, the eID Wallet fulfils these objectives; calls on the Commission to ensure that any age assurance mechanisms put on the market to allow access to online services are not based on tracking or profiling of individuals, and fulfil a maximum level of privacy and accuracy for the verification of the users’ age; stresses the need to ensure that online platforms only receive a yes or no response to a query on whether a user is of a certain age, rather than identifying a user; calls on the Commission to introduce a principle of non-linkability to ensure that neither governments nor online platforms can trace on which service users have used their age assurance tool;

26. Stresses that it is the provider’s primary responsibility to ensure effective age assurance for minors accessing their service; welcomes the Commission’s consideration that it remains the responsibility of the provider to ensure that the method used by a third party is effective;

27. Notes the fact that several providers of online platforms have put in place mechanisms to control and/or verify who can access their services, some of which have serious problems regarding their accuracy and privacy or rely on mechanisms that are biased; notes, furthermore, that some online platforms have already stated in their terms and conditions that their services cannot be used by minors under 13; stresses that despite these mechanisms, minors are, to a large extent, exposed to, actively recommended or targeted with content that is not age-appropriate or are exposed to risks such as addiction to social media, video-sharing platforms or video games; stresses that the current mechanisms can be circumvented easily and their use needs to be complemented by strong and appropriate safeguards to guarantee a sufficient level of protection of minors online;

28. Calls for the establishment of a harmonised European digital age limit of 16 as the default threshold under which access to online social media platforms should not be allowed unless parents or guardians have authorised their children otherwise; calls for the same age limit to apply to video-sharing platforms and AI companions that present risks to minors; calls, furthermore, for a harmonised European digital age limit of 13, under which no minor can access social media platforms; calls on the panel of experts commissioned by President von der Leyen to assess, as part of their mandate, the application of an age limit of 13 for video-sharing services and AI companions; stresses that this would serve as a protective measure to support parents in managing their children’s digital presence and ensuring age-appropriate online engagement;

29. Calls, furthermore, on the Commission to consider introducing personal liability for senior management in cases of serious and persistent non-compliance with the protection provisions concerning minors of Articles 34-35 of the DSA, in order to incentivise better compliance;

30. Underlines that while online platforms roll out age assurance mechanisms, more action is needed to make online services safe for minors in particular, as well as for the rest of society; concludes that addictive features and other problematic mechanisms cannot remain and need to be addressed urgently;

Parental control

31. Calls for greater collaboration between parents, communications providers, media companies, video game publishers and developers, and online platforms when it comes to identifying threats to child safety online and in the exchange of best practice in this area;

32. Stresses that the DSA requires online platforms to ensure that effective safeguards are in place to protect vulnerable groups, such as minors; highlights the fact that providers of online platforms that are accessible to minors can put in place easily understandable and user-friendly parental control measures to allow parents and guardians to help protect minors against the risk of exposure to harmful content; recalls that the AVMSD sets out similar obligations for video-sharing platforms; highlights that parental control measures should be effective and designed with privacy in mind, and should avoid excessive monitoring of minors’ online activities; considers that the use of these tools should be optional for parents; emphasises that parental control tools are without prejudice to the responsibility of online platforms to ensure that their products are safe and appropriate for minors by design or by default, for instance by setting the highest security and privacy settings of accounts held by minors by default, or by designing clear and concise easy-to-find reporting mechanisms, in particular for minors;

33. Expresses concern that parents often lack the knowledge, competences, and adequate and efficient tools to monitor and manage minors’ online activities or are unaware of the tools available to them; notes that even when parental control tools are used, they are not always easy to find or navigate, and minors can easily circumvent them; stresses the importance of ensuring that such tools are user-friendly, intuitive and easily found and understood by all parents and guardians, including those with disabilities; urges platforms to take greater responsibility in promoting and improving the effectiveness of parental control systems, and calls on them to publish relevant data on the effectiveness of these systems;

34. Believes that while parental control tools offer a solution to limit children’s exposure to inappropriate content, services or excessive screen time, they remain fragmented across different platforms and devices, and are often difficult for the average parent or guardian to use effectively; calls on the Commission to encourage online platforms and other digital services to develop interoperable parental control tools, including a basic set of interoperable features, in order to overcome operating system and device fragmentation and enable parents to manage their children’s device and application access regardless of brand or OS; believes that this approach should also be part of the updated Commission guidelines on protecting minors;

35. Underlines the importance of existing public awareness initiatives by certain Member States to support parents and guardians in ensuring the safety of minors on the internet and social media platforms by providing information and awareness on the use of tools for parental control;

Filling the gaps in current EU legislation

Digital addiction

36. Acknowledges that certain design features can also offer some benefits, such as helping users find relevant and high-quality content, and enabling user-driven assessment of content quality; stresses, however, that the addictive use of digital products and services carries the risk of economic, physical and mental harm to minors; highlights the fact that addictive design features are often inherent to the business model of platforms, notably social media, as well as to certain online services and online games, but they are also used by services providers that do not fall within the scope of the DSA, resulting in the commercial exploitation of minors; highlights, furthermore, the fact that some addictive design features could also manifest themselves in algorithmic recommender systems that prioritise emotive, extreme, inappropriate or (hyper)personalised content to maximise user engagement; is concerned that online platforms may push minors into harmful ‘rabbit holes’ of toxic content or take advantage of minors’ vulnerabilities through advertising; notes that the Commission has already launched investigations relative to addictive designs under the DSA; calls for the effective enforcement of existing provisions under the DSA that address addictive design features as part of the risk assessment and mitigation obligations; calls on the Commission to propose legislation that mandates age-appropriate design, and safety by design and by default, requesting that all platforms and other traders include the necessary risk-based safeguards in their recommender systems, ban engagement-based recommender algorithms for minors, ban the most harmful addictive practices and disable other addictive design features by default for minors, where appropriate;

37. Stresses that features such as ‘infinite scrolling’[20], ‘auto play’[21], ‘pull to refresh’, disappearing stories, rewards and incentive features for continuous or repeated use and penalties for inactivity, streaks and excessive push notifications, gambling-like mechanics (loot boxes) and harmful gamification practices are aimed, by design, at influencing minors’ decision-making, drawing them in with manipulative strategies that are aimed at increasing their engagement and the amount of time and money they spend online and can increase addiction; stresses that such features can negatively affect minors’ ability to concentrate and engage healthily with online content; stresses that features such as auto play are already used for video content targeting minors; reiterates Parliament’s call to effectively tackle addictive design in all online services and for all ages in the review of consumer protection law, and urgently calls for ambitious action at EU level in this regard;

38. Stresses that certain features of digital services, such as the ‘like’ mechanism, social validation loops and ‘read receipts’, can be seen as promoting approval-seeking behaviour and social comparison and undermining self-acceptance and self-worth when the desired feedback is not received, and highlights the fact that they are used by online platforms and traders as a powerful motivator for engagement, especially for minors;

39. Calls on the Commission to ensure that providers of social media and video-sharing platforms’ recommender systems do not present content to minors based on profiling; stresses that this helps protect minors’ mental health, reduce addictive behaviours and encourage more meaningful offline experiences;

40. Underlines the mental and physical health effects of excessive screen time and problematic smartphone use; welcomes the announcement by the President of the Commission in her 2025 State of the Union address to convene a panel of experts on children’s access to social media before end of the year; calls on the Commission to urgently convene such a panel and to ensure it works in collaboration with the World Health Organization and the national health authorities of the Member States, as well as youth and parent representatives, in order to issue European evidence-based state-of-the art guidelines on screen time for minors to improve the protection of minors as consumers online, building on the work already done in France, Sweden and the Netherlands; notes that these guidelines should include guidance on age-appropriate screen use, including maximum time per age, recommendations per type of online service broken down by different age categories, a recommended minimum age for the use of smartphones and, where relevant, other devices, and for example recommendations on smartphone use in schools; highlights the urgency of progressing on the findings of the EU-wide inquiry on the broader impacts of social media, which should cover screen time more comprehensively, and may serve as a basis for further action to protect minors; calls for the Commission and the Member States to address the issue of blurring user control over time spent online and excessive screen time and to raise awareness of parents on the effects and risks of excessive screen time, and problematic internet use, where relevant, on the basis of these guidelines;

41. Underlines the growing problem of the online dissemination of extreme content that may be considered illegal content; considers that this is often closely linked to the commercial design of certain online platforms, which may include features that foster digital addiction, maximise engagement and negatively affect civic discourse; notes that minors may be particularly vulnerable to such content due to developmental factors, such as heightened emotional reactivity, lack of critical judgement and limited impulse control; highlights that since minors spend considerable time online, their exposure to extreme content online can progressively contribute to the normalisation of certain forms of extremism; considers that such a growing problem should be addressed urgently as part of the risk assessments and mitigation measures under Articles 34 and 35 of the DSA, which explicitly mention gender-based violence, protection of minors, mental well-being and the importance of content moderation and algorithmic amplification in this regard; notes that there has been a notable rise in extremist activity among minors in recent years;

Dark patterns

42. Stresses that 97 %[22] of the most popular websites and apps used by EU consumers deploy at least one dark pattern; stresses that minors are particularly vulnerable to dark patterns; underlines that because of their developmental stage and psychological vulnerabilities, minors are more easily manipulated into, for example, sharing personal data, making unintended online purchases or installing applications; underlines the need for robust enforcement of the relevant legislation, in particular the UCPD and DSA provisions, to prevent manipulative interface design, particularly when targeting vulnerable users;

43. Stresses that there is no single comprehensive EU regulation on dark patterns; recalls that the EU has taken steps to address dark patterns through various legal instruments, such as the UCPD and the DSA; stresses, however, that the scope of the prohibition of dark patterns in the DSA is limited to online platforms whereas dark patterns are widely used across all sectors; recalls that the UCPD provides a general prohibition of dark patterns and that only certain dark patterns are directly prohibited by the UCPD black-list and that none of the current prohibitions refer specifically to digital interfaces; is concerned that this results in legal uncertainty for consumers, including minors, businesses and authorities;

44. Urges, therefore, the Commission to clarify the interplay between existing legislation regulating dark patterns and to consider reviewing the list of unfair commercial practices under Annex I to the UCPD in order to close further legal gaps where appropriate; points in particular to practices such as giving more prominence to certain choices when asking consumers for a decision, preselecting options, requesting that consumers make a choice that has already been made, hampering free choice, such as by removing all stopping cues, making it difficult to change the default settings, using emotional language to steer consumers towards certain choices, using ambiguous language to confuse consumers when asking them for a decision, and creating a sense of urgency by falsely indicating that other users are currently also interested in certain products or services or that only a few items are left, or falsely indicating that many users recently purchased a particular product;

Video games

45. Highlights the fact that online video games fall within the scope of the DSA insofar as they qualify as an intermediary service or a platform; calls on the Commission to vigorously enforce the relevant provisions of the DSA; stresses, however, that the providers of online video games that do not qualify as intermediary services or platforms are not subject to the rules laid down in the DSA; calls on the Commission, where applicable, to ensure minimum DSA-equivalent obligations for all online video games regardless of whether they qualify as an online platform under the DSA or not; calls on the Commission to require child impact assessments for online games targeted at minors, in order to continuously identify and mitigate risks to children;

46. Highlights the video game industry’s commitment to self-regulation, as demonstrated by the implementation of age recommendation systems, which play a role in safeguarding younger audiences and promoting responsible gaming practices; recalls that all video games supplied in Europe should abide by the PEGI’s code of conduct rules;

47. Expresses concern that video games are an increasingly commercial environment for minors, and often encourage unwanted spending and involve the sale of virtual items, including uncertainty-based rewards and gambling-like mechanisms such as loot boxes, and the use of intermediate in-app virtual currencies, which could make it difficult to assess the real value of the transaction and the real price of each individual purchase;

48. Highlights how video games that feature chat boxes enabling user-on-user communication pose a risk for children and make them vulnerable to being surreptitiously coerced into purchasing items and sharing personal data, including banking details;

49. Regrets that there is no EU legislation that specifically regulates features such as virtual items or in-app currencies; stresses that some Member States have adopted national laws to further regulate gaming and that this leads to market fragmentation, which is detrimental to consumers and EU companies; calls on the Commission to assess these national initiatives in order to determine whether an initiative at EU level is required to avoid market fragmentation and to ensure a harmonised and high level of protection across the Union;

50. Notes that many minors watch e-sports competitions; regrets, aside from legitimate sponsorship of e-sports teams or competitions, the increase in sponsorship of e-sports players, teams or competitions by third-party websites that allow virtual items to be exchanged or bought and sold for real money through the use of automation software (bots), thereby distorting in-game economies and undermining fair competition; recalls that this practice often breaches the terms and conditions of video game publishers; calls on video game publishers to take measures to prevent these practices where there is a breach of the terms and conditions;

51. Stresses that minors with disabilities should be able to play online video games; notes the efforts by video game companies towards accessibility and encourages video game companies to continue their efforts; calls on the Commission, in conjunction with video game publishers, to present initiatives to improve the accessibility of online video games for people with disabilities;

52. Urges the Commission to guarantee a high level of protection for minors who play video games, in particular by prohibiting loot boxes and other randomised content such as wheels of fortune, prize wheels, card packs in exchange for real money in games that are likely to be accessed by minors, as well as to address risks linked to in-app currencies, micro-transactions, pay-to-progress and pay-to-win mechanisms that are accessible to minors and that can be presented in a way that encourages overspending or gambling behaviours or that are embedded in the core gameplay dynamic, and therefore forced on the users;

Toys

53. Highlights that toys incorporating artificial intelligence (AI) may present risks to children’s physical and mental health, as well as to privacy, data protection and security; underlines that, in line with the forthcoming Toy Safety Regulation[23], toys containing AI systems as safety components are considered high-risk AI systems and require third-party conformity assessment unless relevant harmonised standards have been applied; calls on manufacturers to ensure compliance with both the Toy Safety Regulation and the Artificial Intelligence Act (AI Act), implementing effective risk management, transparency, human oversight and robust data governance to safeguard minors online;

Targeted advertising

54. Stresses that tracking and targeting increases the asymmetry between minors on the one hand and traders and digital service providers on the other hand, and increases the risk of unfair commercial practices, notably manipulation; notes that the tracking of minors online can also present a security risk, for example by making them more vulnerable to fraud, manipulation or foreign influence if the data collected about them is used against them; points out that minors are particularly vulnerable in this regard as they can be more easily exploited or deceived; calls, therefore, on the Commission and the Member States to pay particular attention to the persistence of such risks, in particular in upcoming initiatives;

55. Highlights that despite the existing legal protections in the DSA, studies have shown that minors can still be targeted with ads via proxy categories; underlines that research has found that numerous advertising practices were hidden and fraudulent and could therefore deceive minors in particular[24]; calls on the Commission to pay particular attention to such possible circumventions of the UCPD when it comes to the protection of minors when enforcing the relevant provisions under the DSA and to address possible shortcomings;

56. Highlights the fact that minors need more protection from behavioural monitoring, profiling and influencing; stresses that the DSA bans targeted advertising to minors based on profiling on online platforms; calls on the Commission to launch investigations when this ban is violated; stresses that this ban does not currently apply to other traders and digital services that may engage in personalised advertising; stresses that minors should enjoy a high and consistent level of protection regardless of the type of service; calls on the Commission to introduce equivalent obligations that would be applicable to all digital services and traders;

57. Highlights the fact that, while public interest communications, such as the promotion of protection helplines for minors, may use similar channels and techniques as advertising, they do not always constitute advertising, especially if their primary purpose is to inform or benefit the public rather than to achieve commercial gain; notes, however, that the ban on targeted advertising for minors as set out in the DSA may prevent such communications; calls on the Commission to assess this issue and consider how such public interest communications that aim to protect minors can continue to be effectively disseminated;

Influencer marketing

58. Highlights the fact that EU consumer law only partly addresses the issues posed by influencer marketing, and that the promotion of unrealistic lifestyle, dangerous or unhealthy products or services to minors by influencers and the harmful commercial practices involved in influencer marketing remain a reality;

59. Stresses also the concerning use of minors’ images by ‘parent influencers’ or ‘family influencers’ outside of any legal or ethical framework, often in exchange for financial remuneration, raising serious questions about consent, privacy and the commercial exploitation of minors;

60. Highlights the fact that despite social media platforms’ minimum age requirements, minors below this age are active on these platforms and operate as kidfluencers; highlights this widespread underage presence and the growing trend of kidfluencing among minors of all ages;

61. Emphasises that followers of kidfluencers are typically minors themselves; stresses that direct exhortation to minors to buy advertised products or persuade their parents or other adults to buy advertised products for them is highly likely to happen via kidfluencer marketing and is therefore contrary to the UCPD[25]; stresses that kidfluencers’ knowledge of the legal rules governing their operations is doubtful;

62. Highlights the challenges that minors may face as influencers or protagonists of the content generated by another influencer (i.e. sharenting); stresses that this activity makes kidfluencers dependent on platforms, on community feedback such as ‘likes’, on viewership data/ratings and on the pressure to constantly create content for commercial gain;

63. Calls on the Commission to protect minors from commercial exploitation, including by prohibiting platforms from monetising or otherwise providing financial or material incentives for kidfluencing; highlights that certain Member States have extended their existing national legal regimes protecting child actors to influencers who are minors; calls on the Member States to consider establishing equivalent levels of protection of minors by applying a similar regulatory approach framework;

64. Highlights the fact that consumers, including minors, face a lack of transparency about the paid promotion of products by social media influencers and have difficulty distinguishing between commercial and non-commercial content; highlights that a number of Member States have started to implement rules for influencers at national level; underlines the need for an EU-wide transparency and disclosure mechanism for advertising content promoted by influencers; underlines, furthermore, the need to reduce fragmented application and enforcement of existing rules and further strengthen EU consumer rules; points out that transparency obligations should be accompanied by clear rules on how to display paid content to consumers, in particular minors, reflecting growing recognition of the need for further transparency obligations and for greater harmonisation across the single market;

65. Welcomes the Commission’s Influencer Legal Hub, which provides resources and training to help influencers and other relevant parties understand and comply with EU consumer and advertising laws;

66. Calls on the Commission to encourage brands to work together with influencers who can be certified voluntarily to ensure they have the knowledge of legal rules applicable to their activities, giving brands cooperating with certified influencers assurance that their products are marketed in a legally compliant manner and enabling users to place greater trust in influencers;

67. Recalls that the AVMSD provides protection for minors against commercial communications encouraging behaviour prejudicial to a person’s health or safety; notes, however, that the AVMSD applies to influencers insofar as they qualify as media service providers; highlights that not all influencers are therefore covered by the AVMSD; notes, furthermore, that the Commission’s 2021 guidance on the UCPD clarified that influencers can be covered by the UCPD, including the prohibition of misleading and unfair commercial practices; notes, however, that in the 2024 Fitness Check on Digital Fairness, the Commission concluded that there is still legal uncertainty about the applicable rules and that a risk of regulatory fragmentation exists; calls, therefore, on the Commission to further assess the need to address legal uncertainties, including by having a harmonised definition of ‘influencer marketing’, and strengthen, where appropriate, the prohibition of commercial practices that include the promotion by influencers of products, trends, challenges or behaviours that may put minors at risk, including health or financial risk;

68. Expresses concern about the rapid spread of AI-generated content impersonating famous personalities, online content creators and brands, which exploit users’ trust for commercial gain or for the dissemination of disinformation; calls on the Commission to classify scams and false impersonation as part of possible systemic risks under the DSA and AI Act and to consider adopting measures to protect minors from being misled and deceived and allow for recourse to businesses and consumers affected by these scams and false impersonations;

69. Calls on the Commission to adapt EU rules, including by clarifying, where necessary, the responsibilities of platforms, influencers and other actors in the value chain, such as their agencies, and the brands whose products and services are being promoted;

Artificial intelligence

70. Underlines the increasing ease of access to generative AI tools, which are also accessible to minors; highlights the risks regarding transparency, training data, manipulation and AI anthropomorphism, deepfakes, companionship chatbots and AI agents, all of which can result in a distortion of reality and in misinformation, or harm mental health, which in some cases has already led to serious tragedies, including suicides, or led users to make unintended online purchases and share personal data; warns, in particular, against the misuse of children’s images available online, which can be exploited by AI technologies for inappropriate or harmful purposes, including the creation of manipulated or abusive content; is deeply concerned about the fact that companies have started to roll back their efforts in testing AI models[26] with regard to safety and fundamental rights, which puts minors at particular risk when interacting with AI; stresses the importance of the implementation and effective enforcement of the AI Act in this regard;

71. Stresses that AI-powered applications may be collecting and processing children’s data without the informed consent or awareness of the child or their legal guardians, especially in regard to the rights of the child as a data subject;

72. Stresses the potential risks of generative AI programs being used to facilitate child sexual abuse grooming by predators, but also deepfake sexualised material of minors, which is used in cyberbullying by peers, and stresses the need to counter this from the side of services providers and platforms that make these tools available;

73. Recognises the significant threat posed by AI-powered nudity applications, which allow users to generate manipulated images of individuals without their consent, thereby infringing upon personal privacy and dignity; acknowledges the potential for such technologies to facilitate gender-based cyberviolence; calls on the Commission to urgently address the ethical and legal challenges posed by these technologies;

74. Emphasises that minors are particularly vulnerable to addictive design features of interactive AI applications and games that often may be intentionally developed to maximise engagement; highlights that AI-driven recommendation algorithms may lead to excessive screen time and create dependency, thus hindering cognitive, emotional and social development in minors and hampering the development of individual skills and competences;

75. Highlights that the AI Act is an important instrument to address the systemic risks posed by AI systems, especially to public health, safety, public security, fundamental rights or society as a whole[27], and to establish a minimum level of safety and trustworthiness for AI in the EU, including the AI systems used in content generation, recommendation systems and user interaction on online platforms; calls for the swift implementation and enforcement of obligations arising from the AI Act, to ensure safe and trustworthy AI that is also safe for minors; calls on the Commission and the Member States’ authorities to firmly enforce the AI Act provisions addressing manipulative and deceptive chatbots;

76. Stresses the critical importance of the Code of Practice on General-Purpose AI (GPAI) models as a cornerstone for the effective implementation of the AI Act’s systemic risk mitigation obligations, particularly in safeguarding fundamental rights and addressing the risks posed to minors;

77. Notes that AI agents, companions and chatbots for minors are increasingly present online and can pose significant risks to minors’ rights and healthy development; stresses that the Commission should, as part of its ongoing evaluation of consumer legislation, consider additional safeguards, transparency standards and a prohibition of human-like features, to avoid nudging children to make purchases, and to ensure the ability for children to easily disengage from AI chatbots;

78. Calls on the Member States to urgently appoint competent authorities in charge of enforcing bans on prohibited practices and enforce the AI Act in an ambitious and effective manner, in particular but not limited to provisions concerning the protection of minors; highlights that competent authorities should pay particular attention to AI systems, notably those deployed by social media platforms, and investigate whether any AI systems likely to interact with minors exploit the vulnerabilities of minors or distort their behaviour in a manner that causes, or is reasonably likely to cause, that person or another person significant harm, as prohibited by the AI Act;

79. Stresses the importance of supporting research on AI for and with children, as well as of promoting incentives to reduce the digital divide;

Media and digital literacy and education

80. Acknowledges the importance of media and digital literacy in empowering minors to navigate online environments safely and responsibly and to apply critical thinking; recalls that the Member States and the European Union have the competencies to adopt specific measures in this area, including in educational settings; recommends the inclusion of media and digital literacy in national curricula and the development of training programmes for teachers and educators on a regular basis; stresses the necessity of linking media literacy to basic skills, such as reading comprehension, and underlines the importance of AI literacy and digital literacy as prerequisites for creating a safer internet environment and recognising online risks; stresses the need to assess both potential harms and available mitigation tools, while also exploring the positive role that digital environments can play in access to information, education and civic participation; acknowledges the role of public service media in providing diverse, high-quality content for minors to reinforce educational and cultural objectives; recommends that digital platforms promote educational materials and cultural diversity, particularly from European creators and public institutions;

81. Underlines that protecting minors online is a shared responsibility that requires the involvement of parents, guardians, public authorities, platforms and schools; highlights the need for adequate resources, and calls for structured cooperation between education systems, digital platforms and regulatory authorities to deliver high-quality digital and media literacy programmes; calls on the Commission to encourage exchange of best practices and cross-border cooperation to ensure safe and responsible online environments;

82. Stresses the need to provide parents and guardians with adequate training and guidance to help them support their children’s digital experience; calls for EU-funded training and awareness initiatives for parents, guardians, teachers and educators, including on the effects of excessive use of social media, exposure to harmful content and online addiction; stresses that minors are particularly vulnerable to online disinformation, on account of their developmental stage and the evolving nature of their media literacy and critical thinking skills; emphasises the urgent need for age-appropriate awareness campaigns in schools to empower minors to recognise and report online risks; encourages online platforms and digital service providers to adopt voluntary measures and tools that help strengthen the digital skills, critical thinking and digital literacy of minors and support parents or guardians in guiding their online experience; recalls that the most vulnerable children offline are often also the most vulnerable online and highlights the need for targeted support for vulnerable children, including those from low-income backgrounds or with disabilities; encourages age-appropriate literacy support for minors acting as digital content creators, with guidance for teachers and youth workers, and ethical guidelines for children’s images and narratives online;

°

° °

83. Instructs its President to forward this resolution to the Council, the Commission and the governments and parliaments of the Member States.


EXPLANATORY STATEMENT

Online services have a huge potential to improve minors’ access to services and information. However, it is important that minors can benefit fully from the digital environment without being manipulated, exploited, or exposed to risks. Ensuring a high level of online protection for minors in the entire digital ecosystem is essential for creating environments where they can actively engage with, learn from, and benefit from the digital possibilities.

The Digital Services Act (DSA) marked a major step forward in strengthening the EU’s digital rulebook. The DSA provides a robust toolbox to ensure the protection of consumers on online platforms, including minors. Among other provisions, it introduces, a framework for prohibiting targeted advertising based on profiling of minors and requires online platforms to implement appropriate and proportionate measures to ensure a high level of privacy, safety, and security for minors. These rules must now be enforced swiftly and effectively by all Member States and the European Commission. They must be accompanied by guidelines to ensure a high-level protection of minors online.

When it comes to age verification mechanisms, the Rapporteur believes that the current fragmentation in the internal market weakens the protection of minors online, and calls on the Commission to address the issue at EU level. The Rapporteur considers that the primary responsibility to ensure adequate protection of minors online lies with the platforms and digital services providers and it cannot be passed on to the parents or guardians. The Rapporteur acknowledges that some online platforms are deploying age estimation and parental control, to help protecting minors against the risk of exposure to harmful content.

Despite the regulatory achievements, minors continue to be exposed to manipulative design features, exploitative commercial practices, and digital environments that prioritise engagement and monetisation over users’ best interests. Online platforms and digital services systematically deploy algorithmic systems, dark patterns, and addictive design techniques engineered to maximise user engagement and prolong time spent online. These systems prioritise emotive and extreme content to maximise user engagement increasing the risk of minors being pushed into harmful “rabbit holes” of toxic content. Moreover, minors experience adverse mental and physical effects which can include anxiety, low self-esteem, disrupted sleep patterns, reduced concentration, and compulsive online behaviour. The algorithmic systems are deliberately constructed to influence behaviour, especially considering minors are more susceptible to such mechanisms, and the design choices are directly linked to platform business models aiming at commercial exploitation of minors.

Additionally, minors are increasingly active in gaming, often involving unwanted commercial practices including the sale of virtual items and gambling-like mechanisms such as loot boxes, and the use of intermediate in-app virtual currencies. These practices are designed to stimulate prolonged engagement through addictive design features. Even though part of the gaming sector falls under the DSA, it remains largely outside the scope of current EU digital rulebook.

Furthermore, the Rapporteur is of the view that in order to ensure a high level of online protection for minors in the entire digital ecosystem, the Commission should consider reviewing existing consumer legislation and where appropriate present new legislative proposals aiming to close the existing gaps in the regulatory framework. In particular, the Rapporteur strongly believes that a horizontal legislative initiative to prohibit addictive design features is much needed at EU level.

As digital technologies evolve, the EU rules, guidelines and mechanisms need to be robust and future proof and effectively enforced in order to ensure the high level of protection of minors online.


 

ANNEX: DECLARATION OF INPUT

Pursuant to Article 8 of Annex I to the Rules of Procedure, the rapporteur declares that she included in her report input on matters pertaining to the subject of the file that she received, in the preparation of the report, prior to the adoption thereof in committee, from the following interest representatives falling within the scope of the Interinstitutional Agreement on a mandatory transparency register[28], or from the following representatives of public authorities of third countries, including their diplomatic missions and embassies:

1. Interest representatives falling within the scope of the Interinstitutional Agreement on a mandatory transparency register

Meta

Apple

DOT Europe

Child focus

Digitalt Ansvar

Børns Vilkår

Middle Tech Europe

Snap, Inc.

Red Barnet

International Federation of Red Cross and Red Crescent Societies

Computer and Communications Industry Association

Radius

Samsung Electronics Europe

Internet Watch Foundation

TikTok Technology Ltd

Avaaz Foundation

Google

European DIGITAL SME Alliance

Zscaler

AWO agency

BEUC

Danske Pressepublikationers Kollektive Forvaltningsorganisation

5 rights

2. Representatives of public authorities of third countries, including their diplomatic missions and embassies

Ministry of Digitalisation and Public Governance, Norway

The list above is drawn up under the exclusive responsibility of the rapporteur.

Where natural persons are identified in the list by their name, by their function or by both, the rapporteur declares that she has submitted to the natural persons concerned the European Parliament's Data Protection Notice No 484 (https://www.europarl.europa.eu/data-protect/index.do), which sets out the conditions applicable to the processing of their personal data and the rights linked to that processing.

 


OPINION OF THE COMMITTEE ON CULTURE AND EDUCATION (21.10.2025)

for the Committee on the Internal Market and Consumer Protection

on protection of minors online

(2025/2060(INI))

Rapporteur for opinion: Ivaylo Valchev 

 

AMENDMENTS

The Committee on Culture and Education submits the following to the Committee on the Internal Market and Consumer Protection, as the committee responsible:

Amendment  1

Motion for a resolution

Citation -1 (new)

 

Motion for a resolution

Amendment

 

 having regard to Articles 114 and 165 of the Treaty on the Functioning of the European Union,

Amendment  2

Motion for a resolution

Citation -1 a (new)

 

Motion for a resolution

Amendment

 

 having regard to the Charter of Fundamental Rights of the European Union (the Charter),

Amendment  3

Motion for a resolution

Citation -1 b (new)

 

Motion for a resolution

Amendment

 

 having regard to the United Nations Convention on the Rights of the Child, adopted on 20 November 1989,

Amendment  4

Motion for a resolution

Citation -1 c (new)

 

Motion for a resolution

Amendment

 

 having regard to the United Nations Convention on the Rights of the Child, adopted on 20 November 1989, and General comment No. 25 (2021) thereon, on children’s rights in relation to the digital environment,

Amendment  5

Motion for a resolution

Citation 1 a (new)

 

Motion for a resolution

Amendment

 

 having regard to Directive (EU) 2018/1808 of the European Parliament and of the Council of 14 November 2018 amending Directive 2010/13/EU on the coordination of certain provisions laid down by law, regulation or administrative action in Member States concerning the provision of audiovisual media services (Audiovisual Media Services Directive) in view of changing market realities1a (the Audiovisual Media Services Directive),

Amendment  6

Motion for a resolution

Citation 1 b (new)

 

Motion for a resolution

Amendment

 

 having regard to the Council conclusions of 2 May 2025 on the assessment of the legal framework for audiovisual media services and video-sharing platform services1a,

Amendment  7

Motion for a resolution

Citation 1 c (new)

 

Motion for a resolution

Amendment

 

 having regard to the Commission communication of 7 July 2020 entitled ‘Guidelines on the practical application of the essential functionality criterion of the definition of a “video-sharing platform service” under the Audiovisual Media Services Directive’1a,

Amendment  8

Motion for a resolution

Citation 1 d (new)

 

Motion for a resolution

Amendment

 

 having regard to Commission staff working document of 5 January 2024 entitled ‘Reporting on the application of Directive 2010/13/EU “Audiovisual Media Services Directive” as amended by Directive (EU) 2018/1808, for the period 2019-2022’ (SWD(2024)0004),

Amendment  9

Motion for a resolution

Citation 1 e (new)

 

Motion for a resolution

Amendment

 

 having regard to its resolution of 9 May 2023 on the implementation of the revised Audiovisual Media Services Directive1a,

Amendment  10

Motion for a resolution

Citation 1 f (new)

 

Motion for a resolution

Amendment

 

 having regard to the Commission communication of 11 May 2022 entitled ‘A Digital Decade for children and youth: the new European strategy for a better internet for kids (BIK+)’ (COM(2022)0212) and its first evaluation1a,

Amendment  11

Motion for a resolution

Citation 1 g (new)

 

Motion for a resolution

Amendment

 

 having regard to the Commission communication of 9 March 2021 entitled ‘2030 Digital Compass: the European way for the Digital Decade’ (COM(2021)0118),

Amendment  12

Motion for a resolution

Citation 1 h (new)

 

Motion for a resolution

Amendment

 

 having regard to its resolution of 10 November 2022 on esports and video games1a,

Amendment  13

Motion for a resolution

Citation 1 i (new)

 

Motion for a resolution

Amendment

 

 having regard to Regulation (EU) 2016/679 of the European Parliament and of the Council of 27 April 2016 on the protection of natural persons with regard to the processing of personal data and on the free movement of such data, and repealing Directive 95/46/EC (General Data Protection Regulation)1a,

Amendment  14

Motion for a resolution

Citation 6 a (new)

 

Motion for a resolution

Amendment

 

 having regard to the study requested by its Committee on Culture and Education entitled ‘The influence of social media on the development of children and young people’, published in February 20231a,

Amendment  15

Motion for a resolution

Citation 6 b (new)

 

Motion for a resolution

Amendment

 

 having regard to the report of May 2025 by the European Union Drugs Agency entitled ‘Key findings from the 2024 European School Survey Project on Alcohol and Other Drugs (ESPAD)’,

Amendment  16

Motion for a resolution

Recital E a (new)

 

Motion for a resolution

Amendment

 

Ea. whereas the obligations laid down by the Audiovisual Media Services Directive (AVMSD) for video-sharing platforms to protect minors from content that may impair their physical, mental or moral development include, as appropriate, establishing and operating age verification systems for users of those platforms, and providing for parental control systems that are under the control of the end-user with respect to the aforementioned content;

Amendment  17

Motion for a resolution

Recital E b (new)

 

Motion for a resolution

Amendment

 

Eb. whereas the Commission needs to make greater effort to raise awareness about the limitations of self-declared age verification;

Amendment  18

Motion for a resolution

Recital E c (new)

 

Motion for a resolution

Amendment

 

Ec. whereas certain forms of age assurance and verification may risk excluding undocumented or marginalised users and could affect privacy and free expression; whereas measures to protect minors should prioritise youth empowerment;

Amendment  19

Motion for a resolution

Recital E d (new)

 

Motion for a resolution

Amendment

 

Ed. whereas ensuring appropriate protection for young people and children online requires targeted measures and education programmes, not only for minors but also for their teachers, parents and caregivers, focusing on prevention techniques, awareness-raising and media and digital literacy campaigns, and informing parents of tools to limit access to age-inappropriate or harmful content;

Amendment  20

Motion for a resolution

Recital E e (new)

 

Motion for a resolution

Amendment

 

Ee. whereas digital education and media literacy are essential for strengthening children’s resilience to online risks, including disinformation and harmful content, while equipping them with the skills to navigate, contribute to and shape the digital world responsibly and confidently;

Amendment  21

Motion for a resolution

Recital E f (new)

 

Motion for a resolution

Amendment

 

Ef. whereas as children develop their digital literacy and online competencies, their capacity for independent engagement increases, requiring age-appropriate educational support to empower informed choices, critical thinking and safe online participation; whereas minors increasingly act as content creators or ‘child influencers’, sometimes involving commercialisation or early public exposure, which can lead to exploitation without consent, highlighting the need for ethical guidance and educational support;

Amendment  22

Motion for a resolution

Recital E g (new)

 

Motion for a resolution

Amendment

 

Eg. whereas media literacy and content labelling initiatives vary widely across platforms, and citizens, especially minors, engage increasingly with social media, video-sharing or content-sharing platforms and online aggregators, where influencers and personalities shape public discourse;

Amendment  23

Motion for a resolution

Recital E h (new)

 

Motion for a resolution

Amendment

 

Eh. whereas parents, teachers and caregivers often lack adequate support, training and tools to help minors navigate online environments safely and responsibly;

Amendment  24

Motion for a resolution

Recital E i (new)

 

Motion for a resolution

Amendment

 

Ei. whereas prevention and digital literacy should be actively promoted as a public good, integrated into lifelong learning strategies, and guided by expert-led approaches to healthy screen time, including age-appropriate limits, thresholds for device use, and parental presence during device use, for children’s well-being;

Amendment  25

Motion for a resolution

Recital E j (new)

 

Motion for a resolution

Amendment

 

E j. whereas the protection of minors online should not only focus on preventing harm, but should also include proactive exposure to enriching, diverse and culturally relevant content;

Amendment  26

Motion for a resolution

Recital E k (new)

 

Motion for a resolution

Amendment

 

Ek. whereas the promotion of high-quality cultural and educational content represents a positive alternative to potentially harmful content, and European cultural institutions can play a crucial role in providing appropriate and enriching online content for minors;

Amendment  27

Motion for a resolution

Recital E l (new)

 

Motion for a resolution

Amendment

 

El. whereas cooperation with the European creative industries to protect minors online should be championed and self-regulatory measures encouraged; whereas the significance of European digital platforms and video games for Europe’s competitiveness and cultural diversity must be properly acknowledged and safeguarded;

Amendment  28

Motion for a resolution

Recital E m (new)

 

Motion for a resolution

Amendment

 

Em. whereas the freedom of expression and of information, as guaranteed by the AVMSD and the DSA, does not impede the adoption of effective measures for the protection of minors online;

Amendment  29

Motion for a resolution

Recital E n (new)

 

Motion for a resolution

Amendment

 

En. whereas empowering young people and children in the digital world is not only a matter of safety, but also of citizenship; whereas young people and children should not be passive technology consumers, but be actively in control of the technologies they use; whereas media and digital literacy are simultaneously key enablers of children’s active, creative and civic participation online, equipping them with the skills to navigate, contribute to and shape the digital world responsibly and confidently, beyond mere risk prevention;

Amendment  30

Motion for a resolution

Recital E o (new)

 

Motion for a resolution

Amendment

 

Eo. whereas in order to ensure safe online experiences and foster digital empowerment and active participation, it is crucial to promote digital education and enhance the digital skills and competencies, including media and critical literacy, of young people and children, as well as of parents and educators, particularly from vulnerable groups, in line with the AVMSD, the Charter, the UN Convention on the Rights of the Child, the European Declaration on Digital Rights and Principles for the Digital Decade1a, the 2030 Digital Compass and the BIK+ strategy; whereas the first evaluation of the BIK+ strategy highlights emerging risks such as harmful content, online scams, privacy violations, adverse effects on mental health, and the monetisation of minors’ online activities; whereas the evaluation stresses the need to enhance inclusivity and empower young users, while adults raise concerns about cyberbullying, misinformation and excessive screen time, pointing to the need for parents, caregivers and teachers to be better prepared to address the challenges of social media, AI and algorithmic exposure;

Amendment  31

Motion for a resolution

Recital E p (new)

 

Motion for a resolution

Amendment

 

Ep. whereas the AVMSD recognises the protection of minors as a key policy objective of the Union and brings additional value, complementing the DSA, by providing specific rules that apply to video-sharing platform providers of all sizes, to protect minors from programs, user-generated videos and audiovisual commercial communication that may impair their physical, mental or moral development;

Amendment  32

Motion for a resolution

Recital E q (new)

 

Motion for a resolution

Amendment

 

Eq. whereas the AVMSD requires video-sharing platforms to implement a feature allowing users to indicate whether uploaded content contains commercial communications, product placement or sponsorship, and sets out advertising quality requirements, notably banning surreptitious or subliminal audiovisual commercial communications, as well as imposing restrictions on audiovisual commercial communications for alcoholic beverages aimed at minors or communications likely to cause physical, mental or moral harm, including by exploiting minors’ inexperience or credulity or depicting them in dangerous situations;

Amendment  33

Motion for a resolution

Recital E r (new)

 

Motion for a resolution

Amendment

 

Er. whereas the freedom of expression and of information, as guaranteed by the AVMSD and the DSA, does not impede the implementation of effective measures for the protection of minors online;

Amendment  34

Motion for a resolution

Recital E s (new)

 

Motion for a resolution

Amendment

 

Es. whereas the Commission should submit an ex post evaluation of the impact of the AVMSD and its added value, accompanied, where appropriate, by proposals for its review;

Amendment  35

Motion for a resolution

Recital E t (new)

 

Motion for a resolution

Amendment

 

Et. whereas the effectiveness of these rules depends on consistent implementation and enforcement across the Member States, as underlined in the Commission’s 2024 report on the application of the AVMSD;

Amendment  36

Motion for a resolution

Recital E u (new)

 

Motion for a resolution

Amendment

 

Eu. whereas algorithmic transparency and external auditing should be emphasised to enforce AVMSD provisions effectively;

Amendment  37

Motion for a resolution

Recital E v (new)

 

Motion for a resolution

Amendment

 

Ev. whereas the ESPAD report highlights the fact that the online environment is not only a technological but also a social and an emotional space and the report links exposure to harmful content, cyberbullying and excessive screen time with increased anxiety, loneliness and sleep difficulties among 15- and 16-year-olds;

Amendment  38

Motion for a resolution

Recital E w (new)

 

Motion for a resolution

Amendment

 

Ew. whereas excessive screen time and social media use negatively affect minors’ physical, mental and social well-being, interpersonal relations and educational performance, and have been associated with stress, sleep disorders, depression, suicidal ideation, low self-esteem and negative body image;

Amendment  39

Motion for a resolution

Recital E x (new)

 

Motion for a resolution

Amendment

 

Ex. whereas many people, including minors, find it challenging to distinguish between true and false information in online news;

Amendment  40

Motion for a resolution

Paragraph 2

 

Motion for a resolution

Amendment

2. Urges the Commission to swiftly adopt ambitious and pragmatic guidelines on the protection of minors online in accordance with the DSA, and stresses that these guidelines should contain recommendations specifying how online platforms are to implement high levels of privacy, safety and security for minors online;

2. Urges the Commission to ensure swift implementation of the guidelines on the protection of minors online in accordance with the DSA and in compliance with other applicable EU legislation, notably the AVMSD and its implementation at national level; stresses that these guidelines should contain recommendations specifying how online platforms are to implement high levels of privacy, safety and security for minors online;

Amendment  41

Motion for a resolution

Paragraph 6

 

Motion for a resolution

Amendment

6. Recalls that the Commission and the Member States have a shared responsibility when it comes to the enforcement of the DSA; expresses concern that there are significant delays in designating and/or empowering Digital Services Coordinators (DSCs) in several Member States; welcomes the infringement procedures opened against a number of Member States that have failed to designate and/or empower DSCs; urges the Member States that have not yet done so to designate and empower DSCs without delay; calls on the Commission to ensure a harmonised approach to enforcement actions by Member States in order to prevent fragmentation in the internal market;

6. Recalls that the Commission and the Member States have a shared responsibility when it comes to the enforcement of the DSA; expresses concern that there are significant delays in designating and/or empowering Digital Services Coordinators (DSCs) in several Member States; welcomes the infringement procedures opened against a number of Member States that have failed to designate and/or empower DSCs; urges the Member States that have not yet done so to designate and empower DSCs without delay; calls on the Commission to ensure a harmonised approach to enforcement actions by Member States in order to prevent fragmentation in the internal market, taking into account sector-specific rules, particularly those concerning the protection of minors under the AVMSD, as well as the margin of discretion granted to Member States in the transposition of these rules in the light of national specificities;

Amendment  42

Motion for a resolution

Paragraph 7

 

Motion for a resolution

Amendment

7. Highlights the fact that there is currently a fragmented approach to age assurance and age verification across the EU, and that this leads to fragmentation in the internal market; regrets the fact that the different national approaches may result in an uneven level of protection of minors in the EU;

7. Highlights the fact that there is currently a fragmented approach to age assurance and age verification across the EU and across platforms, and that this leads to fragmentation in the internal market; regrets the fact that the different national approaches may result in an uneven level of protection of minors in the EU;

Amendment  43

Motion for a resolution

Paragraph 8

 

Motion for a resolution

Amendment

8. Calls on the Commission to put forward recommendations for effective age assurance or age verification mechanisms to protect minors online, in accordance with the DSA, as a first step, and to present appropriate legislative measures if necessary;

8. Notes the Commission’s work towards an EU approach on privacy-preserving and reliable age verification, including releasing the blueprint on age verification and testing it in a pilot phase, and the recommendations and guidance for effective age assurance or age verification mechanisms in the guidelines to protect minors online, in accordance with the DSA and the AVMSD and in line with children’s rights, as a first step, and calls for the Commission to present appropriate legislative measures if gaps persist, particularly in the protection against risks of overuse, addiction, exposure to harmful content and commercial exploitation;

Amendment  44

Motion for a resolution

Paragraph 9

 

Motion for a resolution

Amendment

9. Highlights the fact that several providers of online platforms have put in place mechanisms to control and/or verify who can access their services; stresses that despite these mechanisms, minors are, to a large extent, viewing content that is not age-appropriate; stresses that the current mechanisms can be circumvented easily and that their use needs to be complemented by strong safeguards to guarantee a sufficient level of protection of minors online;

9. Highlights the fact that several providers of online platforms have put in place mechanisms to control and/or verify who can access their services; stresses that despite these mechanisms, minors are, to a large extent, viewing content, especially audiovisual content, that is not age-appropriate; stresses that the current mechanisms can be circumvented easily and that their use needs to be complemented by strong safeguards to guarantee a sufficient level of protection of minors online; calls for a principled and child-centred approach and prevention strategies to ensure that digital platforms that are accessible to minors uphold the highest standards of safety, privacy and well-being;

Amendment  45

Motion for a resolution

Paragraph 9 a (new)

 

Motion for a resolution

Amendment

 

9a. Welcomes Member States’ initiatives assessing digital maturity and calls on the Commission to examine the possibility of setting an EU-wide digital maturity age;

Amendment  46

Motion for a resolution

Paragraph 10

 

Motion for a resolution

Amendment

10. Stresses that the DSA requires online platforms to ensure that effective safeguards are in place to protect vulnerable groups, such as minors; highlights the fact that providers of online platforms that are accessible to minors may put in place parental control measures to allow parents and guardians to help protect minors against the risk of exposure to harmful content; emphasises that online platforms should not use these tools as a reason to transfer their responsibility to protect minors online to parents;

10. Stresses that the DSA requires online platforms to ensure that effective safeguards are in place to protect vulnerable groups, such as minors; highlights the fact that providers of online platforms that are accessible to minors may put in place parental control measures to allow parents and guardians to help protect minors against the risk of exposure to harmful content; emphasises that online platforms should not use these tools as a means of transferring their responsibility to protect minors online to parents; recalls that the AVMSD sets out similar obligations for video-sharing platforms;

Amendment  47

Motion for a resolution

Paragraph 11

 

Motion for a resolution

Amendment

11. Expresses concern that parents often lack adequate tools to monitor and manage minors’ online activities or are unaware of the tools available to them; notes that even when parental control tools are used, minors can easily circumvent them; stresses the importance of ensuring that such tools are user-friendly, intuitive and easily understood by parents; urges the platforms to take greater responsibility in promoting and improving the effectiveness of parental control systems, and calls on them to publish relevant data on the effectiveness of these systems;

11. Expresses concern that parents often are unaware of the tools available to them to monitor and manage minors’ online activities; notes that even when parental control tools are used, minors can sometimes circumvent them; stresses the importance of ensuring that such tools are user-friendly, intuitive, accessible and easily understood by all parents and caregivers, including those with disabilities; urges the platforms to take greater responsibility in promoting and improving the effectiveness of parental control systems; encourages public authorities and the Member States to collaborate with platforms and other stakeholders to inform and educate parents, caregivers and teachers about parental tools, and calls on them to publish relevant data on the effectiveness of these systems; welcomes initiatives and relevant tools, such as parental control apps, available at national level, to help parents control their children’s online activities;

Amendment  48

Motion for a resolution

Paragraph 11 a (new)

 

Motion for a resolution

Amendment

 

11a. Encourages platforms to adopt age-appropriate communication tools to explain risks and features to minors, their parents and caregivers and to provide support in cooperation with schools; stresses the importance of including parental control tools in platform design and opposes practices that pressure minors into oversharing personal data; calls on platforms, in this regard, to comply with the principles of minimal exposure and data minimisation, to be transparent and to offer routes for challenge and redress;

Amendment  49

Motion for a resolution

Subheading 3 a (new)

 

Motion for a resolution

Amendment

 

Media and digital literacy and education

Amendment  50

Motion for a resolution

Paragraph 11 b (new)

 

Motion for a resolution

Amendment

 

11b. Recognises the central role of media and digital literacy in empowering minors, parents, caregivers, teachers and educators to navigate the digital environment safely and critically; recommends the inclusion of media and digital literacy in national curricula and the development of continuous training programmes for teachers and educators; stresses that media literacy should be linked to basic skills, such as reading comprehension, and underlines the importance of AI literacy covering algorithmic bias, content moderation and data rights; stresses the need to assess both potential harms and available mitigation tools, while also exploring the positive role that digital environments can play in providing access to information, education and civic participation;

Amendment  51

Motion for a resolution

Paragraph 11 c (new)

 

Motion for a resolution

Amendment

 

11c. Stresses the importance of age-appropriate awareness campaigns in schools to help minors identify and report online risks; calls for EU-funded training and awareness initiatives for parents, caregivers, teachers and educators, including on the effects of excessive use of social media, exposure to harmful content, and online addiction; recalls that the most vulnerable children offline are also often the most vulnerable online, particularly those who may lack parental oversight, and highlights the need for targeted support for vulnerable children, including those from low-income backgrounds or with disabilities; advocates age-appropriate literacy support for minors acting as digital content creators, with guidance for teachers and youth workers, and ethical guidelines for children’s images and narratives online;

Amendment  52

Motion for a resolution

Paragraph 11 d (new)

 

Motion for a resolution

Amendment

 

11d. Highlights the need for adequate resources, and calls for structured cooperation between education systems, digital platforms and regulatory authorities to deliver high-quality digital and media literacy programmes; calls on the Commission to encourage exchange of best practice and cross-border cooperation to ensure safe and responsible online environments;

Amendment  53

Motion for a resolution

Paragraph 11 e (new)

 

Motion for a resolution

Amendment

 

11e. Calls on the Commission and the Member States to address the effects on health of excessive screen time, smartphone use and addictive design, through research, awareness campaigns and balanced measures that protect cultural and educational access; welcomes, in this regard, the upcoming EU-wide inquiry into the impact of social media on health and well-being; welcomes initiatives to reduce screen time in schools and recalls the role of parents in safe gaming, while stressing the need to promote a safe gaming environment and encouraging awareness of Pan-European Game Information (PEGI ) ratings and parental controls;

Amendment  54

Motion for a resolution

Subheading 3 b (new)

 

Motion for a resolution

Amendment

 

Culture and cultural preservation

Amendment  55

Motion for a resolution

Paragraph 11 f (new)

 

Motion for a resolution

Amendment

 

11f. Acknowledges the role of public service media in providing diverse, high-quality content for minors, to counter negative trends and reinforce educational and cultural objectives;

Amendment  56

Motion for a resolution

Paragraph 11 g (new)

 

Motion for a resolution

Amendment

 

11g. Supports the creation of safe and interactive digital environments for minors through partnerships with online platforms, museums, libraries and educational institutions, highlighting European cultural heritage;

Amendment  57

Motion for a resolution

Paragraph 11 h (new)

 

Motion for a resolution

Amendment

 

11h. Urges digital platforms to promote cultural diversity and educational materials, particularly from European creators and public institutions;

Amendment  58

Motion for a resolution

Subheading 3 c (new)

 

Motion for a resolution

Amendment

 

Involving young people

Amendment  59

Motion for a resolution

Paragraph 11 i (new)

 

Motion for a resolution

Amendment

 

11i. Stresses the importance of involving, and encouraging close cooperation among, developers, researchers, psychologists, teachers and educators, parents and caregivers, youth organisations, cultural and civil society actors, to ensure the development and implementation of digital spaces and experiences that are safe, inclusive, empowering and supportive of young people’s well-being;

Amendment  60

Motion for a resolution

Paragraph 11 j (new)

 

Motion for a resolution

Amendment

 

11j. Call, furthermore, on the Commission and the Member States to support programmes enabling minors and families to understand manipulative design features and to resist addictive behaviour, in order to strengthen resilience and informed digital use, while ensuring a balanced approach that preserves access to educational and cultural digital content and maintains innovation;

Amendment  61

Motion for a resolution

Subheading 4 a (new)

 

Motion for a resolution

Amendment

 

Implementation of the AVMSD and the BIK+ strategy and overall efficiency and coherence of the EU framework

Amendment  62

Motion for a resolution

Paragraph 11 k (new)

 

Motion for a resolution

Amendment

 

11k. Recalls that the Member States must protect minors from content that impairs their physical, mental or moral development;

Amendment  63

Motion for a resolution

Paragraph 11 l (new)

 

Motion for a resolution

Amendment

 

11l. Stresses the need for coherence between the DSA, the AVMSD and the overarching BIK+ strategy and for enhanced cross-border cooperation to ensure that the regulations are implemented consistently on video-sharing platforms and to address violations in their application, in order to guarantee effective child protection; underlines the need to complement regulatory efforts under the DSA with educational measures at EU and national levels;

Amendment  64

Motion for a resolution

Paragraph 11 m (new)

 

Motion for a resolution

Amendment

 

11m. Calls on platforms to implement swiftly the recommendations of the guidelines on protection of minors under Article 28 DSA, and, in particular, to simplify terms and conditions, provide clear reporting processes and offer guidance for younger users;

Amendment  65

Motion for a resolution

Paragraph 11 n (new)

 

Motion for a resolution

Amendment

 

11n. Calls on the Commission, the European Board for Digital Services and national authorities to monitor how online platforms protect minors, and to consider how current legislation can be better applied to ensure a safer digital environment for children;

Amendment  66

Motion for a resolution

Paragraph 11 o (new)

 

Motion for a resolution

Amendment

 

11o. Notes the need for sufficient resources for national media regulators and EU-level support for smaller Member States;

Amendment  67

Motion for a resolution

Paragraph 11 p (new)

 

Motion for a resolution

Amendment

 

11p. Notes that divergent implementation of the General Data Protection Regulation1a (GDPR) across Member States may undermine efforts to ensure a safe and age-appropriate online environment for children; underlines that the AVMSD sets crucial obligations for video-sharing platforms to protect minors from harmful content and to ensure transparency in advertising; recalls that the consistent application of these provisions is vital for safeguarding the physical, mental and moral development of children online; stresses the need for closer coordination of and support for Member States to guarantee consistent child protection standards, especially for national media regulators and for smaller Member States;

Amendment  68

Motion for a resolution

Subheading 6 a (new)

 

Motion for a resolution

Amendment

 

Disinformation, online radicalisation and cyberbullying

Amendment  69

Motion for a resolution

Paragraph 18 a (new)

 

Motion for a resolution

Amendment

 

18a. Stresses that minors are particularly vulnerable to online disinformation, on account of their developmental stage and the evolving nature of their media literacy and critical thinking skills;

Amendment  70

Motion for a resolution

Paragraph 18 b (new)

 

Motion for a resolution

Amendment

 

18b. Urges the Member States to coordinate strategies between schools, families, youth organisations and platforms, providing teachers with appropriate training on cyberbullying awareness and prevention and providing children with age-appropriate reporting channels and psychological support, building on current practices, in particular within the network of Safer Internet Centres and in civil society organisations;

Amendment  71

Motion for a resolution

Paragraph 18 c (new)

 

Motion for a resolution

Amendment

 

18c. Underlines that cyberbullying and online hate speech, affecting minors in particular, often lead to stigmatisation, exclusion and mental health issues;

Amendment  72

Motion for a resolution

Paragraph 18 d (new)

 

Motion for a resolution

Amendment

 

18d. Calls on the Commission to develop common guidelines and tools, including fact-checking techniques and source analysis, to counter disinformation aimed at minors, including by labelling content and removing harmful online content, false information and hate speech, to enable users to recognise credible sources and to prevent radicalisation;

Amendment  73

Motion for a resolution

Paragraph 18 e (new)

 

Motion for a resolution

Amendment

 

18e. Warns of the dangers posed by minors being targeted by radicalising narratives online and calls for comprehensive and balanced prevention strategies that focus on early warning, critical thinking and digital mentorship;

Amendment  74

Motion for a resolution

Paragraph 18 f (new)

 

Motion for a resolution

Amendment

 

18f. Calls for a European research initiative dedicated to understanding how children and teenagers behave online and how they are exposed to disinformation and harmful content, in order to enhance evidence-based prevention and ensure platform accountability;

Amendment  75

Motion for a resolution

Paragraph 30

 

Motion for a resolution

Amendment

30. Highlights how the AI Act is addressing the systemic risks posed by AI systems, including those used in content generation, recommendation systems and user interaction on online platforms; calls for the swift implementation of the transparency obligations arising from the AI Act, in particular with regard to the effective watermarking of AI-generated content on online platforms;

30. Highlights how the AI Act is addressing the systemic risks posed by AI systems, including those used in content generation, recommendation systems and user interaction on online platforms; calls for the swift implementation of the transparency obligations arising from the AI Act, in particular with regard to the effective detection, labelling and watermarking of AI-generated content on online platforms;


ANNEX: DECLARATION OF INPUT

The rapporteur for opinion declares under his exclusive responsibility that he did not include in his opinion input from interest representatives falling within the scope of the Interinstitutional Agreement on a mandatory transparency register[29], or from representatives of public authorities of third countries, including their diplomatic missions and embassies, to be listed in this Annex pursuant to Article 8 of Annex I to the Rules of Procedure.

 


 

INFORMATION ON ADOPTION BY THE COMMITTEE ASKED FOR OPINION

 

Date adopted

24.9.2025

Result of final vote

+ : 26

- : 0

0 : 1

Members present for the final vote

Nikolaos Anadiotis, Zsuzsanna Borvendég, Laurence Farreng, Gabriela Firea, Sunčana Glavak, Esteban González Pons, Catherine Griset, Hannes Heide, Lara Magoni, Eleonora Meleti, Nikos Pappas, Hristo Petrov, Giusi Princi, Sabrina Repp, Diana Riba i Giner, Nela Riehl, Manuela Ripa, Sandro Ruotolo, Joanna Scheuring-Wielgus, Marco Squarta, Zoltán Tarr, Eugen Tomac, Ivaylo Valchev, Annamária Vicsek, Bogdan Andrzej Zdrojewski, Milan Zver

Substitutes present for the final vote

Maria Guzenina


FINAL VOTE BY ROLL CALL
BY THE COMMITTEE ASKED FOR OPINION

26

+

ECR

Magoni Lara, Squarta Marco, Valchev Ivaylo

ESN

Borvendég Zsuzsanna

NI

Anadiotis Nikolaos

PPE

Glavak Suncana, González Pons Esteban, Meleti Eleonora, Princi Giusi, Ripa Manuela, Tarr Zoltán, Zdrojewski Bogdan Andrzej

PfE

Griset Catherine, Vicsek Annamária

Renew

Farreng Laurence, Petrov Hristo, Tomac Eugen

S&D

Firea Gabriela, Guzenina Maria, Heide Hannes, Repp Sabrina, Ruotolo Sandro, Scheuring-Wielgus Joanna

The Left

Pappas Nikos

Verts/ALE

Riba i Giner Diana, Riehl Nela

 

0

-

 

 

 

1

0

PPE

Zver Milan

 

Key to symbols:

+ : in favour

- : against

0 : abstention

 


 

 

INFORMATION ON ADOPTION IN COMMITTEE RESPONSIBLE

Date adopted

16.10.2025

 

 

 

Result of final vote

+:

–:

0:

32

5

9

Members present for the final vote

Peter Agius, Alex Agius Saliba, Jeannette Baljeu, Laura Ballarín Cereza, Arno Bausemer, Biljana Borzan, Anna Cavazzini, Stefano Cavedagna, David Cormand, Henrik Dahl, Dóra Dávid, Adnan Dibrani, Regina Doherty, Kamila Gasiuk-Pihowicz, Maria Grapini, Elisabeth Grossmann, Maria Guzenina, Svenja Hahn, Anna-Maja Henriksson, Virginie Joron, Kateřina Konečná, Katrin Langensiepen, Piotr Müller, Denis Nesci, Christel Schaldemose, Andreas Schwab, Tomislav Sokol, Dimitris Tsiodras, Inese Vaidere, Adina Vălean, Kim Van Sparrentak, Marion Walsmann

Substitutes present for the final vote

Marc Angel, Jaroslav Bžoch, José Cepeda, Johan Danielsson, Gaetano Pedulla’, Paulius Saudargas, Sabine Verheyen, Stéphanie Yon-Courtin, Kosma Złotowski

Members under Rule 216(7) present for the final vote

Csaba Dömötör, Dan-Ştefan Motreanu, Hristo Petrov, Pascale Piera, Volker Schnurrbusch

 


FINAL VOTE BY ROLL CALL BY THE COMMITTEE RESPONSIBLE

32

+

NI

Kateřina Konečná

PPE

Peter Agius, Dóra Dávid, Regina Doherty, Kamila Gasiuk-Pihowicz, Dan-Ştefan Motreanu, Paulius Saudargas, Andreas Schwab, Tomislav Sokol, Dimitris Tsiodras, Inese Vaidere, Adina Vălean, Sabine Verheyen, Marion Walsmann

Renew

Jeannette Baljeu, Stéphanie Yon-Courtin

S&D

Alex Agius Saliba, Marc Angel, Laura Ballarín Cereza, Biljana Borzan, José Cepeda, Johan Danielsson, Adnan Dibrani, Maria Grapini, Elisabeth Grossmann, Maria Guzenina, Christel Schaldemose

The Left

Gaetano Pedulla'

Verts/ALE

Anna Cavazzini, David Cormand, Katrin Langensiepen, Kim Van Sparrentak

 

5

-

ECR

Piotr Müller, Kosma Złotowski

ESN

Arno Bausemer, Volker Schnurrbusch

Renew

Svenja Hahn

 

9

0

ECR

Stefano Cavedagna, Denis Nesci

PPE

Henrik Dahl

PfE

Jaroslav Bžoch, Csaba Dömötör, Virginie Joron, Pascale Piera

Renew

Anna-Maja Henriksson, Hristo Petrov

 

Key to symbols:

+ : in favour

- : against

0 : abstention

 

 

 

Last updated: 12 November 2025
Legal notice - Privacy policy