European Parliament resolution of 26 November 2025 on the protection of minors online (2025/2060(INI))
The European Parliament,
– having regard to Regulation (EU) 2022/2065 of the European Parliament and of the Council of 19 October 2022 on a Single Market For Digital Services and amending Directive 2000/31/EC (Digital Services Act)(1),
– having regard to its resolution of 12 December 2023 on addictive design of online services and consumer protection in the EU single market(2),
– having regard to its resolution of 12 March 2009 on the protection of consumers, in particular minors, in respect of the use of video games(3),
– having regard to its resolution of 18 January 2023 on consumer protection in online video games: a European Single Market approach(4),
– having regard to the UN Convention on the Rights of the Child of 20 November 1989 and to General Comment No. 25 (2021) of the UN Committee on the Rights of the Child of 2 March 2021 on children’s rights in relation to the digital environment,
– having regard to Regulation (EU) 2024/1689 of the European Parliament and of the Council of 13 June 2024 laying down harmonised rules on artificial intelligence and amending Regulations (EC) No 300/2008, (EU) No 167/2013, (EU) No 168/2013, (EU) 2018/858, (EU) 2018/1139 and (EU) 2019/2144 and Directives 2014/90/EU, (EU) 2016/797 and (EU) 2020/1828 (Artificial Intelligence Act)(5),
– having regard to Directive 2005/29/EC of the European Parliament and of the Council of 11 May 2005 concerning unfair business-to-consumer commercial practices in the internal market and amending Council Directive 84/450/EEC, Directives 97/7/EC, 98/27/EC and 2002/65/EC of the European Parliament and of the Council and Regulation (EC) No 2006/2004 of the European Parliament and of the Council (‘Unfair Commercial Practices Directive’)(6),
– having regard to the Commission’s fitness check of EU consumer law on digital fairness(7),
– having regard to the Commission notice of 17 December 2021 entitled ‘Guidance on the interpretation and application of Directive 2005/29/EC of the European Parliament and of the Council concerning unfair business-to-consumer commercial practices in the internal market(8) (the UCPD Guidance),
– having regard to the Commission communication of 7 October 2025 entitled ‘Guidelines on measures to ensure a high level of privacy, safety and security for minors online, pursuant to Article 28(4) of Regulation (EU) 2022/2065’(9),
– having regard to Regulation (EU) 2017/2394 of the European Parliament and of the Council of 12 December 2017 on cooperation between national authorities responsible for the enforcement of consumer protection laws and repealing Regulation (EC) No 2006/2004(10) (Consumer Protection Cooperation Regulation),
– having regard to Directive (EU) 2018/1808 of the European Parliament and of the Council of 14 November 2018 amending Directive 2010/13/EU on the coordination of certain provisions laid down by law, regulation or administrative action in Member States concerning the provision of audiovisual media services (Audiovisual Media Services Directive) in view of changing market realities(11),
– having regard to its resolution of 9 May 2023 on the implementation of the revised Audiovisual Media Services Directive(12),
– having regard to the Commission communication of 11 May 2022 entitled ‘A Digital Decade for children and youth: the new European strategy for a better internet for kids (BIK+)’ (COM(2022)0212),
– having regard to Regulation (EU) 2024/1183 of the European Parliament and of the Council of 11 April 2024 amending Regulation (EU) No 910/2014, as regards establishing the European Digital Identity Framework(13) (eIDAS2 Regulation),
– having regard to the Commission proposal of 6 February 2024 for a directive of the European Parliament and of the Council on combating child sexual abuse and sexual exploitation of children and child sexual abuse material and replacing Council Framework Decision 2004/68/JHA (COM(2024)0060),
– having regard to Regulation (EU) 2016/679 of the European Parliament and of the Council of 27 April 2016 on the protection of natural persons with regard to the processing of personal data and on the free movement of such data, and repealing Directive 95/46/EC (General Data Protection Regulation)(14),
– having regard to Rule 55 of its Rules of Procedure,
– having regard to the opinion of the Committee on Culture and Education,
– having regard to the report of the Committee on the Internal Market and Consumer Protection (A10-0213/2025),
A. whereas 97 % of young people use the internet daily; whereas 78 % of 13 to 17-year-olds report checking their devices at least hourly and 46 % report checking them almost constantly(15); whereas 16 to 24-year-olds spend an average of over seven hours a day on the internet(16); whereas more than half of EU consumers regularly play video games and among children aged from 11 to 14, that share is as high as 84 %(17); whereas among minors that play video games, 64 % of them spend an average of EUR 1-20 per month on games, while there has been an overall average monthly spending increase by children and consumers in general, from EUR 33 in 2020 to EUR 39 in 2023(18);
B. whereas one in four children and young people display ‘problematic’ or ‘dysfunctional’ smartphone use, meaning behavioural patterns mirroring addiction(19); whereas minors are more vulnerable to AI-generated content, AI companions and chatbots, as their cognitive abilities are still developing; whereas research shows that habitual checking of social media in teenagers may be associated with changes in neural sensitivity as regards anticipation of social rewards and punishments; whereas research suggests that the rise in mental health problems in adolescents may be related to excessive social media use, and social media pressure has been identified as one of the top five causes of mental health difficulties for minors, such as heightened anxiety and depression; whereas the World Health Organization recognises gaming addiction as a mental health disorder; whereas studies have shown that excessive social media can impact brain development and brain capacity of minors; whereas correlations have been found between excessive social media use and lack of impulse control;
C. whereas citizens, especially minors, are increasingly engaging with social media, video- and content-sharing platforms, and online aggregators, including for news consumption; whereas public figures and influencers active on these online platforms have gained considerable influence in shaping public discourse; whereas many users, including minors, find it challenging to distinguish between true and false information in online news; whereas ensuring the responsible operation and design of these platforms, alongside the promotion of critical thinking and media and digital literacy among minors, must be a priority;
D. whereas digital education and media literacy are essential for strengthening the resilience of minors to online risks, including disinformation and harmful content, while equipping them with the skills to navigate, contribute to and shape the digital world responsibly and confidently; whereas comprehensive educational programmes and awareness campaigns are essential to inform minors, parents, guardians and educators about risks and safe practices online; whereas this includes targeted measures for minors, parents, guardians and educators, prevention techniques, awareness-raising and media and digital literacy campaigns, and information on tools to limit access to age-inappropriate or harmful content;
E. whereas digital technologies and online services can offer benefits for minors by opening up new opportunities for performance, creativity and expression and supporting education; whereas it is essential to guarantee minors’ right of access of to the internet as it contributes to their freedom of expression, learning, information and opinion; whereas minors should not be passive technology consumers, but be actively in control of the technologies they use; whereas media and digital literacy are simultaneously key enablers of children’s active, creative and civic participation online, equipping them with the competences to actively take part in the digital world in a responsible and confident manner, beyond mere risk prevention;
F. whereas the influencer advertising market in Europe has witnessed significant growth in recent years, driven by the increasing importance of social media platforms, a rise in digital advertising budgets and the increasing preference of brands for more authentic forms of marketing; whereas as businesses strive to engage with younger, tech-savvy consumers, influencer marketing has emerged as a key strategy for building brand awareness, trust and credibility; whereas the European market for influencer marketing is projected to grow at a compound annual growth rate of around 12 % over the next five years, and by 2030 the market size is expected to surpass EUR 6,4 billion, demonstrating robust growth prospects; whereas only 20 % of influencers systematically indicate the commercial nature of the content they share;
G. whereas more and more minors are becoming content creators or kidfluencers, sometimes involving commercialisation or early public exposure, which can lead to exploitation without consent;
H. whereas at present, no EU legislation comprehensively addresses the protection of minors across the digital ecosystems; whereas minors should enjoy equal levels of protection regardless of where they live; whereas online platforms and applications have different purposes and services with various designs, interfaces and features, which can lead to different risks to minors, and, therefore, proportionality and risk-based assessment should be applied; whereas minors should enjoy a higher level of protection online, which requires a harmonised EU approach; whereas ensuring higher levels of protection of minors needs to be balanced with respecting their right to access information and participate in online communities;
I. whereas the Digital Services Act (DSA) recognises the protection of minors as an important policy objective of the Union and includes key provisions to ensure better protection of minors online, such as Article 14(3), Article 28, Article 34(1)(d), Article 35(1)(j) and Article 44(1)(j); whereas the Commission guidelines on measures to ensure a high level of privacy, safety and security for minors online, pursuant to Article 28(4) of Regulation (EU) 2022/2065, further clarify the rules on the protection of minors online;
J. whereas the Audiovisual Media Services Directive (AVMSD) identifies the protection of minors as a key policy objective and lays down provisions to strengthen their online safety, notably on video-sharing platforms;
K. whereas the Pan-European Game Information (PEGI) system, as a European video game content rating system, can help consumers – especially parents – make informed decisions about the suitability of video games and apps for different age groups;
L. whereas evidence from several countries in the EU shows that minors have access to pornographic content, sometimes involuntarily, at a very young age; whereas the viewing of pornography can expose children to violent and age-inappropriate content, which has been proven to increase aggression and sexual violence and may trigger harmful sexual behaviours among children and teenagers;
M. whereas the DSA recognises that parental control tools are among risk mitigation measures for online platforms accessible to minors, as provided for in Article 35(1)(j), and thus play a role in managing the risks to minors’ safety, privacy and well-being in the digital environment;
N. whereas the number of young women that are subject to gender-based violence is increasing year by year, in particular in the online sphere; whereas the lack of action and the persistence of online abuse pushes young women out of the digital sphere and impacts their freedom of expression;
O. whereas there is an urgent need to effectively implement and enforce current EU legislation and to address outstanding challenges in order to create a safe online environment for all users, in particular minors; whereas collaboration with Member States is necessary to ensure consistent implementation and enforcement of protective measures across the EU; whereas the enforcement of existing EU legislation remains under-resourced across Member States, resulting in a fragmented implementation and disparities regarding online protection needed for minors; whereas international cooperation with non-EU countries and global organisations is essential to address the cross-border nature of online risks and harms;
P. whereas several Member States have called for a common approach on digital majority and have already adopted national laws in this respect;
Q. whereas the protection of minors’ data is of utmost importance, including ensuring that their data is not misused or exploited, and that a high level of minors’ privacy, safety and security is safeguarded in all online interactions, in line with the DSA;
R. whereas the freedom of expression and of information, as guaranteed by the DSA, does not impede the implementation of effective measures for the protection of minors online;
S. whereas the Commission has committed, in its mission letters and hearings, to proposing a digital fairness act that would tackle unethical techniques and commercial practices related to dark patterns, marketing by social media influencers, the addictive design of digital products and services, and online profiling, especially when consumer vulnerabilities are exploited for commercial purposes for all consumers, including minors;
Implementation and enforcement of existing legislation
1. Stresses that the DSA requires online platforms that are accessible to minors to put in place appropriate and proportionate measures to ensure that minors benefit from a high level of privacy, safety and security when using their services;
2. Takes note of the Commission guidelines on measures to ensure a high level of privacy, safety and security for minors online, pursuant to Article 28(4) of the DSA; welcomes the broad scope of the guidelines, which include principles such as privacy and security by design and practical guidance on the application of assurance mechanisms following a risk-based approach, including the recommendation for high-risk platforms to adopt accurate, robust and privacy-preserving age verification mechanisms; welcomes, furthermore, the guidelines’ support for protective default settings, the reaffirmed responsibility of app providers, regardless of delegation to operating system (OS) operators, and the promotion of existing tools such as the online safety-by-design codes and Child Rights Impact Assessments;
3. Considers that the Commission guidelines could be further improved, in particular regarding the protection of minors from addictive design, profiling-based recommendations and engagement-based recommendations; notes, furthermore, that these guidelines are not legally binding and that further legislative action might be required;
4. Stresses the importance of the swift implementation and effective enforcement of the DSA and other relevant legislation, including the AVMSD, and its full implementation at national level, in order to achieve the objective of guaranteeing a safe online environment for minors; welcomes the Commission’s investigations under the DSA in the field of the protection of minors online, which focus, in particular, on addictive behaviour, the ‘rabbit hole’ phenomenon, age assurance mechanisms and recommender systems; urges the Commission to conclude its investigations imminently and to take any additional steps needed, including fines and effective corrective measures, in accordance with the rules established under the DSA;
5. Recalls that the Commission, on 27 May 2025, initiated formal proceedings against major pornographic platforms for breaches of obligations arising from the DSA; emphasises that the Commission’s investigations specifically focus on the protection of minors online and the lack of effective measures to verify users’ age; recalls that pornographic content, as well as violent content and content that perpetuates misogynistic, racist or homophobic views, can cause extremely serious physical, social and emotional consequences for minors, with harmful effects on their psychological development;
6. Stresses the importance of coordinated actions of the Consumer Protection Cooperation (CPC) Network in the field of the protection of minors online, which focus, in particular, on direct appeals to children in advertisements, the use of aggressive commercial practices, the lack of clear and transparent information and influencer marketing;
7. Expresses concern about the spread of dangerous trends and challenges on online platforms, including targeted disinformation campaigns, noting that these could harm minors; highlights the lack of awareness among parents and guardians, who often have limited capacity to react before harm is done; urges online platforms to proactively assess risks for minors, in line with the existing legislation, and detect emerging dangerous trends and challenges; calls on the Commission to establish, in cooperation with the Safer Internet Centres, an efficient rapid alert mechanism at EU level to detect such issues and to require platforms to mitigate these issues; calls on the Commission to propose that platforms display crisis resource panels with contact numbers for third-party support services, such as eating disorder or suicide prevention charities;
8. Expresses concern about the recruitment of minors by criminal networks on online platforms; expresses concern about the lack of more ambitious mitigation measures to protect minors, in particular when it comes to notice and action, hyper-personalised and engagement-based recommendation algorithms leading to addictive behaviour, and dark patterns; notes with alarm the recent trend by some large online platforms indicating a relaxation of strict content moderation practices on their platforms; expresses deep concern that such a move may increase the exposure of minors and other vulnerable users to harmful, illegal or inappropriate content, thereby undermining efforts to ensure a safe online environment;
9. Highlights the importance of risk assessment provisions in the DSA and stresses that the risk assessments carried out by very large online platforms and very large online search engines under the DSA with regard to the safety of minors online are often inadequate, do not go into sufficient detail beyond publicly available information, lack proper independent oversight and underestimate the risks posed to minors; calls on the Commission to make full use of the tools available under the DSA to address these issues and to make sure that obligations regarding risk assessments and related mitigation measures are fully respected; calls on the Commission, as part of its enforcement powers under the DSA, to also verify that online platforms are complying with their own terms and conditions; calls on the Commission to consider, as a last resort, banning persistently non-compliant sites or applications; calls on the Commission to provide guidance, criteria, benchmarks and a clear framework for independent audits;
10. Calls on the Commission to make risk assessments more robust by specifying the information, data and assessments that must be included in systemic risk assessments, such as revenue data from personalised versus non-personalised ads for children and an assessment of any risk to minors and the effect of algorithms on minors’ user experience, in particular how algorithms may expose minors to illegal content or addictive design features; stresses the importance of including minors, guardians and all relevant experts in risk assessments in order to ensure that the best interests and needs of minors are duly taken into consideration when designing the service; expresses major concerns about the ongoing failure of major digital platforms to adequately protect minors on their services, as exposed by whistle-blowers who have revealed serious breaches and harmful practices; calls on the Commission to actively engage with whistle-blowers who have disclosed concerns related to the protection of minors on digital platforms, in order to benefit from their expertise and insights with a view to improving the effectiveness of EU policies and enforcement mechanisms aimed at safeguarding minors online;
11. Recalls that the Commission and the Member States have a shared responsibility when it comes to the enforcement of the DSA; expresses concern that there are significant delays in designating and/or empowering Digital Services Coordinators (DSCs) in several Member States, which significantly undermines the protection of consumers, including minors, across the EU; welcomes the infringement procedures opened against a number of Member States that have failed to designate and/or empower DSCs; urges the Member States that have not yet done so to designate and empower DSCs without delay and to allocate adequate resources to DSCs and other relevant authorities involved in DSA enforcement; stresses the need to ensure, in the multiannual financial framework, that authorities have all the necessary resources to carry out their work properly and effectively;
12. Calls on the Commission to ensure a harmonised approach to enforcement actions by Member States, including through harmonised operational procedures for DSCs, in order to prevent fragmentation in the internal market; invites the Commission, the Board for Digital Services Coordinators and the relevant national authorities to closely monitor and evaluate the effectiveness of measures taken by online platforms to protect minors; urges the Commission to facilitate cooperation between DSCs and CPC authorities at Member State level and provide guidance to help Member States to structure this cooperation;
13. Welcomes the enforcement efforts made by the CPC Network in the field of consumer protection, notably e-commerce and video games, to protect minors from harmful practices, as well as the continued dialogue with the video game sector; urges the Commission to propose a revision of the CPC Regulation to ensure that the Commission has centralised investigative and enforcement powers across the EU in cases of widespread infringement of EU consumer laws;
14. Welcomes the Commission’s ‘digital fairness’ fitness check of consumer law; underlines that some issues remain unaddressed concerning the protection of consumers online, with minors identified as particularly vulnerable to unfair and manipulative online practices; calls on the Commission to strengthen the protection of minors online through the upcoming digital fairness act; believes that persuasive technologies used by online actors, such as targeted advertising, influencer marketing, addictive design, loot boxes, in-app currencies in video games, and dark patterns, should fall under the digital fairness act, which should close legal loopholes and be consistent with current legal instruments in order to better protect consumers online, in particular minors, taking into account the need to avoid unnecessary regulatory burdens;
Age verification
15. Takes note of the fact that there is currently a fragmented approach to age assurance (age verification, age estimation and self-declaration) measures across the EU, and that this leads to fragmentation in the internal market; acknowledges that some Member States have implemented advanced measures to enhance the protection of minors online; regrets that the different national approaches may result in an uneven level of protection of minors online across the EU; acknowledges the importance of the ongoing discussions at both national and EU level on this issue, with a view to determining how to best address this challenge at EU level;
16. Underlines that cultural norms, societal values and public sensitivities regarding digital tools and data use vary significantly across Member States, which influences how societies perceive, accept and trust measures such as age verification; stresses that these differences must be carefully considered to ensure that any European-level initiatives respect national contexts and public trust, while providing effective tools to protect minors;
17. Welcomes the Commission’s risk-based approach taken in its guidelines for effective, proportionate and privacy-preserving age assurance mechanisms to protect minors online, in accordance with the DSA; highlights that the Commission considers the use of access restrictions supported by age verification methods an appropriate and proportionate measure to ensure a high level of privacy, safety and security for minors; calls on the Commission to present, if necessary, appropriate legislative measures to provide legal certainty and ensure a harmonised approach for safe and reliable age assurance mechanisms, taking into account the Commission guidelines on the protection of minors throughout the EU, and guaranteeing an equal level of protection of minors in the Union;
18. Notes, however, the concerns related to age assurance mechanisms regarding possible unintended consequences against the freedoms and rights of the child; calls, therefore, for any legislative measures related to age assurance mechanisms to be thoroughly assessed for their impact on fundamental rights, to avoid enshrining surveillance practices, and to ensure that any solutions proposed prioritise the most effective and privacy-preserving measures and are adopted only where strictly necessary, proportionate and accompanied by robust safeguards;
19. Calls for the Commission to ensure consistency in the area of minors’ safety online across the various pieces of EU legislation in this area;
20. Stresses that age assurance mechanisms are not a silver bullet for protecting minors from the many risks they face online; emphasises that such tools cannot remove platforms’ accountability for ensuring their products are safe and appropriate for minors by design and by default; encourages the development of versions of social media services that are age-appropriate for minors;
21. Highlights the initiative of Commission President Ursula von der Leyen to create a panel of experts to swiftly assess, among other issues, the question of digital age limits and which online services should be covered;
22. Considers that, when requiring confirmation of users’ age, the Commission should only require the use of the most accurate, robust and privacy-preserving mechanisms; considers that such mechanisms should be systemically implemented for online platforms or digital services selling products or services that fall under an age restriction under EU or national legislation and for social media platforms, video-sharing platforms and AI companions, which present risk to minors;
23. Encourages collaboration between the Member States that have already developed digital tools for effective age verification and the Commission in order to develop a strong and effective age verification tool at EU level, ensuring full respect for the technical specifications derived from the eIDAS2 Regulation;
24. Welcomes the Commission’s efforts to provide an EU-wide solution by developing an age-verification app and the EU Digital Identity (eID) Wallet; recommends that the Commission further invest in privacy-preserving systems, such as zero-knowledge proof protocols, building on the principles of minimal exposure and data minimisation; calls on the Commission to ensure that age assurance mechanisms that are deployed are accurate, effective, reliable, robust, non-intrusive, privacy-preserving and non-discriminatory by design, and do not incentivise the development of technologies that rely on the surveillance of users, in particular minors, nor on the deployment of other unethical technological practices;
25. Emphasises that individuals should be granted access to online platforms that use the easiest, safest and most privacy-friendly option by default; considers that, to date, the eID Wallet fulfils these objectives; calls on the Commission to ensure that any age assurance mechanisms put on the market to allow access to online services are not based on tracking or profiling of individuals, and fulfil a maximum level of privacy and accuracy for the verification of the users’ age; stresses the need to ensure that online platforms only receive a yes or no response to a query on whether a user is of a certain age, rather than identifying a user; calls on the Commission to introduce a principle of non-linkability to ensure that neither governments nor online platforms can trace on which service users have used their age assurance tool;
26. Stresses that it is the provider’s primary responsibility to ensure effective age assurance for minors accessing their service; welcomes the Commission’s consideration that it remains the responsibility of the provider to ensure that the method used by a third party is effective;
27. Notes the fact that several providers of online platforms have put in place mechanisms to control and/or verify who can access their services, some of which have serious problems regarding their accuracy and privacy or rely on mechanisms that are biased; notes, furthermore, that some online platforms have already stated in their terms and conditions that their services cannot be used by minors under 13; stresses that despite these mechanisms, minors are, to a large extent, exposed to, actively recommended or targeted with content that is not age-appropriate or are exposed to risks such as addiction to social media, video-sharing platforms or video games; stresses that the current mechanisms can be circumvented easily and their use needs to be complemented by strong and appropriate safeguards to guarantee a sufficient level of protection of minors online;
28. Calls for the establishment of a harmonised European digital age limit of 16 as the default threshold under which access to online social media platforms should not be allowed unless parents or guardians have authorised their children otherwise; calls for the same age limit to apply to video-sharing platforms and AI companions that present risks to minors; calls, furthermore, for a harmonised European digital age limit of 13, under which no minor can access social media platforms; calls on the panel of experts commissioned by President von der Leyen to assess, as part of their mandate, the application of an age limit of 13 for video-sharing services and AI companions; stresses that this would serve as a protective measure to support parents in managing their children’s digital presence and ensuring age-appropriate online engagement;
29. Calls, furthermore, on the Commission to consider introducing personal liability for senior management in cases of serious and persistent non-compliance with the protection provisions concerning minors of Articles 34-35 of the DSA, in order to incentivise better compliance;
30. Underlines that while online platforms roll out age assurance mechanisms, more action is needed to make online services safe for minors in particular, as well as for the rest of society; concludes that addictive features and other problematic mechanisms cannot remain and need to be addressed urgently;
Parental control
31. Calls for greater collaboration between parents, communications providers, media companies, video game publishers and developers, and online platforms when it comes to identifying threats to child safety online and in the exchange of best practice in this area;
32. Stresses that the DSA requires online platforms to ensure that effective safeguards are in place to protect vulnerable groups, such as minors; highlights the fact that providers of online platforms that are accessible to minors can put in place easily understandable and user-friendly parental control measures to allow parents and guardians to help protect minors against the risk of exposure to harmful content; recalls that the AVMSD sets out similar obligations for video-sharing platforms; highlights that parental control measures should be effective and designed with privacy in mind, and should avoid excessive monitoring of minors’ online activities; considers that the use of these tools should be optional for parents; emphasises that parental control tools are without prejudice to the responsibility of online platforms to ensure that their products are safe and appropriate for minors by design or by default, for instance by setting the highest security and privacy settings of accounts held by minors by default, or by designing clear and concise easy-to-find reporting mechanisms, in particular for minors;
33. Expresses concern that parents often lack the knowledge, competences, and adequate and efficient tools to monitor and manage minors’ online activities or are unaware of the tools available to them; notes that even when parental control tools are used, they are not always easy to find or navigate, and minors can easily circumvent them; stresses the importance of ensuring that such tools are user-friendly, intuitive and easily found and understood by all parents and guardians, including those with disabilities; urges platforms to take greater responsibility in promoting and improving the effectiveness of parental control systems, and calls on them to publish relevant data on the effectiveness of these systems;
34. Believes that while parental control tools offer a solution to limit children’s exposure to inappropriate content, services or excessive screen time, they remain fragmented across different platforms and devices, and are often difficult for the average parent or guardian to use effectively; calls on the Commission to encourage online platforms and other digital services to develop interoperable parental control tools, including a basic set of interoperable features, in order to overcome operating system and device fragmentation and enable parents to manage their children’s device and application access regardless of brand or OS; believes that this approach should also be part of the updated Commission guidelines on protecting minors;
35. Underlines the importance of existing public awareness initiatives by certain Member States to support parents and guardians in ensuring the safety of minors on the internet and social media platforms by providing information and awareness on the use of tools for parental control;
Filling the gaps in current EU legislation
Digital addiction
36. Acknowledges that certain design features can also offer some benefits, such as helping users find relevant and high-quality content, and enabling user-driven assessment of content quality; stresses, however, that the addictive use of digital products and services carries the risk of economic, physical and mental harm to minors; highlights the fact that addictive design features are often inherent to the business model of platforms, notably social media, as well as to certain online services and online games, but they are also used by services providers that do not fall within the scope of the DSA, resulting in the commercial exploitation of minors; highlights, furthermore, the fact that some addictive design features could also manifest themselves in algorithmic recommender systems that prioritise emotive, extreme, inappropriate or (hyper)personalised content to maximise user engagement; is concerned that online platforms may push minors into harmful ‘rabbit holes’ of toxic content or take advantage of minors’ vulnerabilities through advertising; notes that the Commission has already launched investigations relative to addictive designs under the DSA; calls for the effective enforcement of existing provisions under the DSA that address addictive design features as part of the risk assessment and mitigation obligations; calls on the Commission to propose legislation that mandates age-appropriate design, and safety by design and by default, requesting that all platforms and other traders include the necessary risk-based safeguards in their recommender systems, ban engagement-based recommender algorithms for minors, ban the most harmful addictive practices and disable other addictive design features by default for minors, where appropriate;
37. Stresses that features such as ‘infinite scrolling’(20), ‘auto play’(21), ‘pull to refresh’, disappearing stories, rewards and incentive features for continuous or repeated use and penalties for inactivity, streaks and excessive push notifications, gambling-like mechanics (loot boxes) and harmful gamification practices are aimed, by design, at influencing minors’ decision-making, drawing them in with manipulative strategies that are aimed at increasing their engagement and the amount of time and money they spend online and can increase addiction; stresses that such features can negatively affect minors’ ability to concentrate and engage healthily with online content; stresses that features such as auto play are already used for video content targeting minors; reiterates Parliament’s call to effectively tackle addictive design in all online services and for all ages in the review of consumer protection law, and urgently calls for ambitious action at EU level in this regard;
38. Stresses that certain features of digital services, such as the ‘like’ mechanism, social validation loops and ‘read receipts’, can be seen as promoting approval-seeking behaviour and social comparison and undermining self-acceptance and self-worth when the desired feedback is not received, and highlights the fact that they are used by online platforms and traders as a powerful motivator for engagement, especially for minors;
39. Calls on the Commission to ensure that providers of social media and video-sharing platforms’ recommender systems do not present content to minors based on profiling; stresses that this helps protect minors’ mental health, reduce addictive behaviours and encourage more meaningful offline experiences;
40. Underlines the mental and physical health effects of excessive screen time and problematic smartphone use; welcomes the announcement by the President of the Commission in her 2025 State of the Union address to convene a panel of experts on children’s access to social media before end of the year; calls on the Commission to urgently convene such a panel and to ensure it works in collaboration with the World Health Organization and the national health authorities of the Member States, as well as youth and parent representatives, in order to issue European evidence-based state-of-the art guidelines on screen time for minors to improve the protection of minors as consumers online, building on the work already done in France, Sweden and the Netherlands; notes that these guidelines should include guidance on age-appropriate screen use, including maximum time per age, recommendations per type of online service broken down by different age categories, a recommended minimum age for the use of smartphones and, where relevant, other devices, and for example recommendations on smartphone use in schools; highlights the urgency of progressing on the findings of the EU-wide inquiry on the broader impacts of social media, which should cover screen time more comprehensively, and may serve as a basis for further action to protect minors; calls for the Commission and the Member States to address the issue of blurring user control over time spent online and excessive screen time and to raise awareness of parents on the effects and risks of excessive screen time, and problematic internet use, where relevant, on the basis of these guidelines;
41. Underlines the growing problem of the online dissemination of extreme content that may be considered illegal content; considers that this is often closely linked to the commercial design of certain online platforms, which may include features that foster digital addiction, maximise engagement and negatively affect civic discourse; notes that minors may be particularly vulnerable to such content due to developmental factors, such as heightened emotional reactivity, lack of critical judgement and limited impulse control; highlights that since minors spend considerable time online, their exposure to extreme content online can progressively contribute to the normalisation of certain forms of extremism; considers that such a growing problem should be addressed urgently as part of the risk assessments and mitigation measures under Articles 34 and 35 of the DSA, which explicitly mention gender-based violence, protection of minors, mental well-being and the importance of content moderation and algorithmic amplification in this regard; notes that there has been a notable rise in extremist activity among minors in recent years;
Dark patterns
42. Stresses that 97 %(22) of the most popular websites and apps used by EU consumers deploy at least one dark pattern; stresses that minors are particularly vulnerable to dark patterns; underlines that because of their developmental stage and psychological vulnerabilities, minors are more easily manipulated into, for example, sharing personal data, making unintended online purchases or installing applications; underlines the need for robust enforcement of the relevant legislation, in particular the UCPD and DSA provisions, to prevent manipulative interface design, particularly when targeting vulnerable users;
43. Stresses that there is no single comprehensive EU regulation on dark patterns; recalls that the EU has taken steps to address dark patterns through various legal instruments, such as the UCPD and the DSA; stresses, however, that the scope of the prohibition of dark patterns in the DSA is limited to online platforms whereas dark patterns are widely used across all sectors; recalls that the UCPD provides a general prohibition of dark patterns and that only certain dark patterns are directly prohibited by the UCPD black-list and that none of the current prohibitions refer specifically to digital interfaces; is concerned that this results in legal uncertainty for consumers, including minors, businesses and authorities;
44. Urges, therefore, the Commission to clarify the interplay between existing legislation regulating dark patterns and to consider reviewing the list of unfair commercial practices under Annex I to the UCPD in order to close further legal gaps where appropriate; points in particular to practices such as giving more prominence to certain choices when asking consumers for a decision, preselecting options, requesting that consumers make a choice that has already been made, hampering free choice, such as by removing all stopping cues, making it difficult to change the default settings, using emotional language to steer consumers towards certain choices, using ambiguous language to confuse consumers when asking them for a decision, and creating a sense of urgency by falsely indicating that other users are currently also interested in certain products or services or that only a few items are left, or falsely indicating that many users recently purchased a particular product;
Video games
45. Highlights the fact that online video games fall within the scope of the DSA insofar as they qualify as an intermediary service or a platform; calls on the Commission to vigorously enforce the relevant provisions of the DSA; stresses, however, that the providers of online video games that do not qualify as intermediary services or platforms are not subject to the rules laid down in the DSA; calls on the Commission, where applicable, to ensure minimum DSA-equivalent obligations for all online video games regardless of whether they qualify as an online platform under the DSA or not; calls on the Commission to require child impact assessments for online games targeted at minors, in order to continuously identify and mitigate risks to children;
46. Highlights the video game industry’s commitment to self-regulation, as demonstrated by the implementation of age recommendation systems, which play a role in safeguarding younger audiences and promoting responsible gaming practices; recalls that all video games supplied in Europe should abide by the PEGI’s code of conduct rules;
47. Expresses concern that video games are an increasingly commercial environment for minors, and often encourage unwanted spending and involve the sale of virtual items, including uncertainty-based rewards and gambling-like mechanisms such as loot boxes, and the use of intermediate in-app virtual currencies, which could make it difficult to assess the real value of the transaction and the real price of each individual purchase;
48. Highlights how video games that feature chat boxes enabling user-on-user communication pose a risk for children and make them vulnerable to being surreptitiously coerced into purchasing items and sharing personal data, including banking details;
49. Regrets that there is no EU legislation that specifically regulates features such as virtual items or in-app currencies; stresses that some Member States have adopted national laws to further regulate gaming and that this leads to market fragmentation, which is detrimental to consumers and EU companies; calls on the Commission to assess these national initiatives in order to determine whether an initiative at EU level is required to avoid market fragmentation and to ensure a harmonised and high level of protection across the Union;
50. Notes that many minors watch e-sports competitions; regrets, aside from legitimate sponsorship of e-sports teams or competitions, the increase in sponsorship of e-sports players, teams or competitions by third-party websites that allow virtual items to be exchanged or bought and sold for real money through the use of automation software (bots), thereby distorting in-game economies and undermining fair competition; recalls that this practice often breaches the terms and conditions of video game publishers; calls on video game publishers to take measures to prevent these practices where there is a breach of the terms and conditions;
51. Stresses that minors with disabilities should be able to play online video games; notes the efforts by video game companies towards accessibility and encourages video game companies to continue their efforts; calls on the Commission, in conjunction with video game publishers, to present initiatives to improve the accessibility of online video games for people with disabilities;
52. Urges the Commission to guarantee a high level of protection for minors who play video games, in particular by prohibiting loot boxes and other randomised content such as wheels of fortune, prize wheels, card packs in exchange for real money in games that are likely to be accessed by minors, as well as to address risks linked to in-app currencies, micro-transactions, pay-to-progress and pay-to-win mechanisms that are accessible to minors and that can be presented in a way that encourages overspending or gambling behaviours or that are embedded in the core gameplay dynamic, and therefore forced on the users;
Toys
53. Highlights that toys incorporating artificial intelligence (AI) may present risks to children’s physical and mental health, as well as to privacy, data protection and security; underlines that, in line with the forthcoming Toy Safety Regulation(23), toys containing AI systems as safety components are considered high-risk AI systems and require third-party conformity assessment unless relevant harmonised standards have been applied; calls on manufacturers to ensure compliance with both the Toy Safety Regulation and the Artificial Intelligence Act (AI Act), implementing effective risk management, transparency, human oversight and robust data governance to safeguard minors online;
Targeted advertising
54. Stresses that tracking and targeting increases the asymmetry between minors on the one hand and traders and digital service providers on the other hand, and increases the risk of unfair commercial practices, notably manipulation; notes that the tracking of minors online can also present a security risk, for example by making them more vulnerable to fraud, manipulation or foreign influence if the data collected about them is used against them; points out that minors are particularly vulnerable in this regard as they can be more easily exploited or deceived; calls, therefore, on the Commission and the Member States to pay particular attention to the persistence of such risks, in particular in upcoming initiatives;
55. Highlights that despite the existing legal protections in the DSA, studies have shown that minors can still be targeted with ads via proxy categories; underlines that research has found that numerous advertising practices were hidden and fraudulent and could therefore deceive minors in particular(24); calls on the Commission to pay particular attention to such possible circumventions of the UCPD when it comes to the protection of minors when enforcing the relevant provisions under the DSA and to address possible shortcomings;
56. Highlights the fact that minors need more protection from behavioural monitoring, profiling and influencing; stresses that the DSA bans targeted advertising to minors based on profiling on online platforms; calls on the Commission to launch investigations when this ban is violated; stresses that this ban does not currently apply to other traders and digital services that may engage in personalised advertising; stresses that minors should enjoy a high and consistent level of protection regardless of the type of service; calls on the Commission to introduce equivalent obligations that would be applicable to all digital services and traders;
57. Highlights the fact that, while public interest communications, such as the promotion of protection helplines for minors, may use similar channels and techniques as advertising, they do not always constitute advertising, especially if their primary purpose is to inform or benefit the public rather than to achieve commercial gain; notes, however, that the ban on targeted advertising for minors as set out in the DSA may prevent such communications; calls on the Commission to assess this issue and consider how such public interest communications that aim to protect minors can continue to be effectively disseminated;
Influencer marketing
58. Highlights the fact that EU consumer law only partly addresses the issues posed by influencer marketing, and that the promotion of unrealistic lifestyle, dangerous or unhealthy products or services to minors by influencers and the harmful commercial practices involved in influencer marketing remain a reality;
59. Stresses also the concerning use of minors’ images by ‘parent influencers’ or ‘family influencers’ outside of any legal or ethical framework, often in exchange for financial remuneration, raising serious questions about consent, privacy and the commercial exploitation of minors;
60. Highlights the fact that despite social media platforms’ minimum age requirements, minors below this age are active on these platforms and operate as kidfluencers; highlights this widespread underage presence and the growing trend of kidfluencing among minors of all ages;
61. Emphasises that followers of kidfluencers are typically minors themselves; stresses that direct exhortation to minors to buy advertised products or persuade their parents or other adults to buy advertised products for them is highly likely to happen via kidfluencer marketing and is therefore contrary to the UCPD(25); stresses that kidfluencers’ knowledge of the legal rules governing their operations is doubtful;
62. Highlights the challenges that minors may face as influencers or protagonists of the content generated by another influencer (i.e. sharenting); stresses that this activity makes kidfluencers dependent on platforms, on community feedback such as ‘likes’, on viewership data/ratings and on the pressure to constantly create content for commercial gain;
63. Calls on the Commission to protect minors from commercial exploitation, including by prohibiting platforms from monetising or otherwise providing financial or material incentives for kidfluencing; highlights that certain Member States have extended their existing national legal regimes protecting child actors to influencers who are minors; calls on the Member States to consider establishing equivalent levels of protection of minors by applying a similar regulatory approach framework;
64. Highlights the fact that consumers, including minors, face a lack of transparency about the paid promotion of products by social media influencers and have difficulty distinguishing between commercial and non-commercial content; highlights that a number of Member States have started to implement rules for influencers at national level; underlines the need for an EU-wide transparency and disclosure mechanism for advertising content promoted by influencers; underlines, furthermore, the need to reduce fragmented application and enforcement of existing rules and further strengthen EU consumer rules; points out that transparency obligations should be accompanied by clear rules on how to display paid content to consumers, in particular minors, reflecting growing recognition of the need for further transparency obligations and for greater harmonisation across the single market;
65. Welcomes the Commission’s Influencer Legal Hub, which provides resources and training to help influencers and other relevant parties understand and comply with EU consumer and advertising laws;
66. Calls on the Commission to encourage brands to work together with influencers who can be certified voluntarily to ensure they have the knowledge of legal rules applicable to their activities, giving brands cooperating with certified influencers assurance that their products are marketed in a legally compliant manner and enabling users to place greater trust in influencers;
67. Recalls that the AVMSD provides protection for minors against commercial communications encouraging behaviour prejudicial to a person’s health or safety; notes, however, that the AVMSD applies to influencers insofar as they qualify as media service providers; highlights that not all influencers are therefore covered by the AVMSD; notes, furthermore, that the Commission’s 2021 guidance on the UCPD clarified that influencers can be covered by the UCPD, including the prohibition of misleading and unfair commercial practices; notes, however, that in the 2024 Fitness Check on Digital Fairness, the Commission concluded that there is still legal uncertainty about the applicable rules and that a risk of regulatory fragmentation exists; calls, therefore, on the Commission to further assess the need to address legal uncertainties, including by having a harmonised definition of ‘influencer marketing’, and strengthen, where appropriate, the prohibition of commercial practices that include the promotion by influencers of products, trends, challenges or behaviours that may put minors at risk, including health or financial risk;
68. Expresses concern about the rapid spread of AI-generated content impersonating famous personalities, online content creators and brands, which exploit users’ trust for commercial gain or for the dissemination of disinformation; calls on the Commission to classify scams and false impersonation as part of possible systemic risks under the DSA and AI Act and to consider adopting measures to protect minors from being misled and deceived and allow for recourse to businesses and consumers affected by these scams and false impersonations;
69. Calls on the Commission to adapt EU rules, including by clarifying, where necessary, the responsibilities of platforms, influencers and other actors in the value chain, such as their agencies, and the brands whose products and services are being promoted;
Artificial intelligence
70. Underlines the increasing ease of access to generative AI tools, which are also accessible to minors; highlights the risks regarding transparency, training data, manipulation and AI anthropomorphism, deepfakes, companionship chatbots and AI agents, all of which can result in a distortion of reality and in misinformation, or harm mental health, which in some cases has already led to serious tragedies, including suicides, or led users to make unintended online purchases and share personal data; warns, in particular, against the misuse of children’s images available online, which can be exploited by AI technologies for inappropriate or harmful purposes, including the creation of manipulated or abusive content; is deeply concerned about the fact that companies have started to roll back their efforts in testing AI models(26) with regard to safety and fundamental rights, which puts minors at particular risk when interacting with AI; stresses the importance of the implementation and effective enforcement of the AI Act in this regard;
71. Stresses that AI-powered applications may be collecting and processing children’s data without the informed consent or awareness of the child or their legal guardians, especially in regard to the rights of the child as a data subject;
72. Stresses the potential risks of generative AI programs being used to facilitate child sexual abuse grooming by predators, but also deepfake sexualised material of minors, which is used in cyberbullying by peers, and stresses the need to counter this from the side of services providers and platforms that make these tools available;
73. Recognises the significant threat posed by AI-powered nudity applications, which allow users to generate manipulated images of individuals without their consent, thereby infringing upon personal privacy and dignity; acknowledges the potential for such technologies to facilitate gender-based cyberviolence; calls on the Commission to urgently address the ethical and legal challenges posed by these technologies;
74. Emphasises that minors are particularly vulnerable to addictive design features of interactive AI applications and games that often may be intentionally developed to maximise engagement; highlights that AI-driven recommendation algorithms may lead to excessive screen time and create dependency, thus hindering cognitive, emotional and social development in minors and hampering the development of individual skills and competences;
75. Highlights that the AI Act is an important instrument to address the systemic risks posed by AI systems, especially to public health, safety, public security, fundamental rights or society as a whole(27), and to establish a minimum level of safety and trustworthiness for AI in the EU, including the AI systems used in content generation, recommendation systems and user interaction on online platforms; calls for the swift implementation and enforcement of obligations arising from the AI Act, to ensure safe and trustworthy AI that is also safe for minors; calls on the Commission and the Member States’ authorities to firmly enforce the AI Act provisions addressing manipulative and deceptive chatbots;
76. Stresses the critical importance of the Code of Practice on General-Purpose AI (GPAI) models as a cornerstone for the effective implementation of the AI Act’s systemic risk mitigation obligations, particularly in safeguarding fundamental rights and addressing the risks posed to minors;
77. Notes that AI agents, companions and chatbots for minors are increasingly present online and can pose significant risks to minors’ rights and healthy development; stresses that the Commission should, as part of its ongoing evaluation of consumer legislation, consider additional safeguards, transparency standards and a prohibition of human-like features, to avoid nudging children to make purchases, and to ensure the ability for children to easily disengage from AI chatbots;
78. Calls on the Member States to urgently appoint competent authorities in charge of enforcing bans on prohibited practices and enforce the AI Act in an ambitious and effective manner, in particular but not limited to provisions concerning the protection of minors; highlights that competent authorities should pay particular attention to AI systems, notably those deployed by social media platforms, and investigate whether any AI systems likely to interact with minors exploit the vulnerabilities of minors or distort their behaviour in a manner that causes, or is reasonably likely to cause, that person or another person significant harm, as prohibited by the AI Act;
79. Stresses the importance of supporting research on AI for and with children, as well as of promoting incentives to reduce the digital divide;
Media and digital literacy and education
80. Acknowledges the importance of media and digital literacy in empowering minors to navigate online environments safely and responsibly and to apply critical thinking; recalls that the Member States and the European Union have the competencies to adopt specific measures in this area, including in educational settings; recommends the inclusion of media and digital literacy in national curricula and the development of training programmes for teachers and educators on a regular basis; stresses the necessity of linking media literacy to basic skills, such as reading comprehension, and underlines the importance of AI literacy and digital literacy as prerequisites for creating a safer internet environment and recognising online risks; stresses the need to assess both potential harms and available mitigation tools, while also exploring the positive role that digital environments can play in access to information, education and civic participation; acknowledges the role of public service media in providing diverse, high-quality content for minors to reinforce educational and cultural objectives; recommends that digital platforms promote educational materials and cultural diversity, particularly from European creators and public institutions;
81. Underlines that protecting minors online is a shared responsibility that requires the involvement of parents, guardians, public authorities, platforms and schools; highlights the need for adequate resources, and calls for structured cooperation between education systems, digital platforms and regulatory authorities to deliver high-quality digital and media literacy programmes; calls on the Commission to encourage exchange of best practices and cross-border cooperation to ensure safe and responsible online environments;
82. Stresses the need to provide parents and guardians with adequate training and guidance to help them support their children’s digital experience; calls for EU-funded training and awareness initiatives for parents, guardians, teachers and educators, including on the effects of excessive use of social media, exposure to harmful content and online addiction; stresses that minors are particularly vulnerable to online disinformation, on account of their developmental stage and the evolving nature of their media literacy and critical thinking skills; emphasises the urgent need for age-appropriate awareness campaigns in schools to empower minors to recognise and report online risks; encourages online platforms and digital service providers to adopt voluntary measures and tools that help strengthen the digital skills, critical thinking and digital literacy of minors and support parents or guardians in guiding their online experience; recalls that the most vulnerable children offline are often also the most vulnerable online and highlights the need for targeted support for vulnerable children, including those from low-income backgrounds or with disabilities; encourages age-appropriate literacy support for minors acting as digital content creators, with guidance for teachers and youth workers, and ethical guidelines for children’s images and narratives online;
o o o
83. Instructs its President to forward this resolution to the Council, the Commission and the governments and parliaments of the Member States.
Maza, M, T., Fox, K., A., Kwon, S., J., Flannery, J., E., Lindquist, K., A., et al., ‘Association of Habitual Checking Behaviors on Social Media With Longitudinal Functional Brain Development’, JAMA Network, (2023), doi:10.1001/jamapediatrics.2022.4924.
Statista, ‘Average daily time spent using the internet by 3rd quarter 2024, by age and gender’, 22 May 2023, https://www.statista.com/statistics/1378510/daily-time-spent-online-worldwide-by-age-and-gender/.
The European Consumer Organisation, ‘Consumer groups denounce video games’ manipulative spending tactics’, 2024, https://www.beuc.eu/press-releases/consumer-groups-denounce-video-games-manipulative-spending-tactics.
The European Consumer Organisation, ‘Monetising Play – Regulating in-Game and in-APP Premium currencies’, 2024, https://www.beuc.eu/sites/default/files/publications/BEUC-X-2024-061_Monetising_play_Regulating_in_game_and_in_app_premium_currencies.pdf.
Lopez-Fernandez, O. and Kuss, D., ‘Harmful Internet Use Part I: Internet addiction and problematic use’, study carried out at the request of the Scientific Foresight Unit of the European Parliamentary Research Service, 31 January 2019, p. 51.
European Commission: Directorate-General for Justice and Consumers, Lupiáñez-Villanueva, F. et al. ‘Behavioural study on unfair commercial practices in the digital environment – Dark patterns and manipulative personalisation – Final report’, Publications Office of the European Union, 2022.
Commission proposal of 28 July 2023 for a regulation of the European Parliament and of the Council on the safety of toys and repealing Directive 2009/48/EC (COM(2023)0462), provisionally agreed between Parliament and Council in April 2025.