The curtain has been drawn on Meta’s movie rating-inspired content filters for teens on Instagram after the Motion Picture Association, which created the ratings, essentially said—loudly and legally—that it didn’t want its ratings anywhere near the social media company’s products.
The clash between Meta and the MPA began in October 2025, when Meta rolled out new content filters on its Teen Accounts for minors on Instagram. According to the company, the new filters would curate age-appropriate content for users guided by PG-13 movie ratings. It appeared to be a way to take advantage of the credibility of the film industry’s existing age ratings, with the rollout accompanied by promotional influencer posts and ads likening the experience to teens watching a movie.
That didn’t fly with the MPA, the film industry group that introduced the PG-13 standard roughly four decades ago. It was squarely unhappy with Meta’s new content filters and sent a cease and desist letter saying that the labeling was “literally false and highly misleading.”
On Tuesday, Meta finally gave up their bid following months of tension. The tech giant and the MPA announced in a joint statement that Meta would “substantially reduce” references to the PG-13 standard in its descriptions and would include a disclaimer when it does so.
“There are lots of differences between social media and movies. We didn’t work with the MPA when updating our content settings, and they’re not rating any content on Instagram, and they’re not endorsing or approving our content settings in any way,” the disclaimer will read once the agreement takes effect on April 15. “Rather, we drew inspiration from the MPA’s public guidelines, which are already familiar to parents. Our content moderation systems are not the same as a movie ratings board, so the experience may not be exactly the same.”
PG-13 movies include some “mature themes” while still keeping it moderate enough that it’s appropriate for teens. A good example would be most Marvel movies, which involve some violence and romantic scenes but nothing too excessive. Meta says that because of the obvious differences between a movie that is two hours long and a photo carousel or a 15-second Reel, they couldn’t issue the PG-13 restrictions in its entirety and instead took inspiration from that rating in deciding what kind of content should be available for teens using the Teen Accounts.
A Meta spokesperson in the statement said that while how the company presents the restrictions in Teen Accounts will change, the criteria that they use will stay the same. The MPA, meanwhile, wants it to be clear once and for all that these two systems are not the same.
“While we welcome efforts to protect kids from content that may not be appropriate for them, this agreement helps ensure that parents do not conflate the two systems – which operate in very different contexts,” MPA CEO and chairman Charles Rivkin said in the release. “The MPA is proud of the trust we have built with parents for nearly sixty years with our film rating system, and we will continue to do everything we can to protect that trust.”
There might be some reason for MPA’s reluctance to be associated with Meta’s social media controls for minors. Namely, Meta’s reputation for child safety has been through the wringer the past couple of months.
Earlier this month, the tech giant was handed losses in two separate landmark cases, one in New Mexico and the other in California, regarding the lack of safety measures for minors on the company’s social media platform, Instagram.
The New Mexico trial claimed that Meta executives created a breeding ground on Instagram for sexual predators looking to exploit children. The California case claimed that deliberate design choices on Instagram, like infinite scrolling, got kids addicted to social media from an early age and worsened serious mental health issues like depression, anxiety, body dysmorphia, and thoughts of self-harm.
“Meta executives knew their products harmed children, disregarded warnings from their own employees, and lied to the public about what they knew,” New Mexico Attorney General Raul Torrez said in a statement following the verdict.
The social media litigation also came on the heels of a Reuters investigation from August 2025, which generated significant backlash with its claims that Meta permitted its AI chatbots to have “sensual” conversations with children.
