Techdirt's think tank, the Copia Institute, is working with the Trust & Safety Professional Association and its sister organization, the Trust & Safety Foundation, to produce an ongoing series of case studies about content moderation decisions. These case studies are presented in a neutral fashion, not aiming to criticize or applaud any particular decision, but to highlight the many different challenges that content moderators face and the tradeoffs they result in. Find more case studies here on Techdirt and on the TSF website.

Content Moderation Case Study: Automated Copyright Takedown Bot Goes Haywire (2018)

from the take-it-all-down dept

Summary: For years, Google and YouTube have included a trusted flagger program by which certain entities that have shown they ?are particularly effective at notifying YouTube? of content violations are given more powerful tools with which to do so.

This is used often in the copyright context, and companies with a good history may be given access to things like bulk flagging tools and priority review of flagged content. One such trusted flagger for copyright was a company called Topple Track, which offered an automated service for musicians, searching the internet for infringing works and dashing off automated DMCA notices.

In May of 2015, digital music distribution company Symphonic purchased Topple Track, but appeared to keep the service running under its own brand.

In the summer of 2018, some people noticed that Topple Track?s automated DMCA notices appeared to go a bit haywire, sending DMCA notices for all kinds of perfectly legitimate content. Among those targeted with DMCA notices were the Electronic Frontier Foundation (EFF), the American Bar Association, NYU?s Law Review, the Crunchbase article about the company MP3Tunes and many, many more — including many artists? own web stores. EFF?s summary of the wild takedowns gives a sample

Among others, these notices improperly target:

Other targets include an article about the DMCA in the NYU Law Review, an NBC News article about anti-virus scams, a Variety article about the Drake-Pusha T feud, and the lyrics to ?Happier? at Ed Sheeran?s official website. It goes on and on.

EFF published an article about this and noted that it seemed as yet another example of an automated DMCA reporting bot ?running amok.? The group also questioned why such a company was in Google?s ?trusted flagger? program.

Decisions to be made by Google / YouTube:

  • What qualifications are there for a partner to be considered a ?trusted flagger??

  • How often are trusted flaggers reviewed to make sure they still belong in the program?

  • What does it take to get a trusted flagger removed from the program?

Questions and policy implications to consider:

  • With more emphasis on the speed of removals, it is often tempting for regulators to promote ?trusted flagging? or ?priority? accounts that are able to get content removed at a much quicker pace. What are the benefits and risks of such programs?

  • Automated flagging and now AI/Machine Learning flagging are increasingly a part of the content moderation landscape. How are they calibrated? How frequently are they reviewed?

  • What should the response be when an automated bot is flagging many accounts mistakenly?

Resolution: After the EFF published its article about Topple Track, the parent company Symphonic Distribution apologized to the organization, saying: ?bugs within the system that resulted in many whitelisted domains receiving these notices unintentionally.? As EFF pointed out in response, this seemed difficult to believe, seeing as the problem was not simply a mistake in domains that shouldn?t have been scanned, but simply claiming stuff that had nothing to do with the underlying copyright-covered material.

A few weeks after the article, YouTube also told EFF that Topple Track had been removed from its Trusted Flagger program, ?due to a pattern of problematic notices.? Some time after this, Topple Track, as a unique organization appeared to disappear, and the service and technology have apparently been subsumed into Symphonic Distribution?s catalog of services.

Originally published on the Trust & Safety Foundation website.

Filed Under: , , , ,
Companies: google, symphonic, topple track, youtube

Rate this comment as insightful
Rate this comment as funny
You have rated this comment as insightful
You have rated this comment as funny
Flag this comment as abusive/trolling/spam
You have flagged this comment
The first word has already been claimed
The last word has already been claimed
Insightful Lightbulb icon Funny Laughing icon Abusive/trolling/spam Flag icon Insightful badge Lightbulb icon Funny badge Laughing icon Comments icon

Comments on “Content Moderation Case Study: Automated Copyright Takedown Bot Goes Haywire (2018)”

Subscribe: RSS Leave a comment
This comment has been deemed funny by the community.
Anonymous Coward says:

Re: How to write a bot

The estate of Janet Jackson and the lawyers for Bread, and for Davido are lining up to send DMCA notices to this blog because you have referred to the title of their songs.

… you might want to be skeptical of the ones saying they representing Jackson’s estate, they might be a little over-eager…

That One Guy (profile) says:

'You're a known liar, so...'

For years, Google and YouTube have included a trusted flagger program by which certain entities that have shown they “are particularly effective at notifying YouTube” of content violations are given more powerful tools with which to do so.

What’s in dire need is a program, or even better a change to the law to ever so slightly balance it out, that allows companies to create and use a list of untrustworthy DMCA filers. If someone sends out bogus DMCA claims left and right there should be a penalty for that, even if it’s only having any of their claims assumed to be bogus until proven otherwise or given the absolute lowest priority in resolving, as without a penalty there’s no incentive for accuracy, something which is demonstrated time and time again.

That Anonymous Coward (profile) says:

Re: Re: Re:

Well many of the sites aren’t in the US and well US law isn’t global.

A DMCA notice is meant to be sent to the hoster of the infringing content, Google does not host the content.

They are abusing the law by demanding a 3rd party with no connection take action to protect their content when the law tells them how to handle it but that’s to hard.

Darkness Of Course (profile) says:

It's just software, how hard can it be?

Machine Learning is a nightmare. There isn’t any review, the test cases are never complete enough. Indeed, they actually cannot be by definition.

Now, add in some oddly connected corporate rat bait and they just tweak their software a little bit. You know, to get a few more hits. It is very easy to do with ML. It is nearly impossible to do it correct without a well defined implementation plan.

We don’t need no stinkin’ plan, said every Corpo Rat ever.

This comment has been flagged by the community. Click here to show it.

Add Your Comment

Your email address will not be published.

Have a Techdirt Account? Sign in now. Want one? Register here

Comment Options:

Make this the or (get credits or sign in to see balance) what's this?

What's this?

Techdirt community members with Techdirt Credits can spotlight a comment as either the "First Word" or "Last Word" on a particular comment thread. Credits can be purchased at the Techdirt Insider Shop »

Follow Techdirt

Techdirt Daily Newsletter

Techdirt Deals
Techdirt Insider Discord
The latest chatter on the Techdirt Insider Discord channel...
Older Stuff
15:43 Content Moderation Case Study: Facebook Struggles To Correctly Moderate The Word 'Hoe' (2021) (21)
15:32 Content Moderation Case Study: Linkedin Blocks Access To Journalist Profiles In China (2021) (1)
16:12 Content Moderation Case Studies: Snapchat Disables GIPHY Integration After Racist 'Sticker' Is Discovered (2018) (11)
15:30 Content Moderation Case Study: Tumblr's Approach To Adult Content (2013) (5)
15:41 Content Moderation Case Study: Twitter's Self-Deleting Tweets Feature Creates New Moderation Problems (2)
15:47 Content Moderation Case Studies: Coca Cola Realizes Custom Bottle Labels Involve Moderation Issues (2021) (14)
15:28 Content Moderation Case Study: Bing Search Results Erases Images Of 'Tank Man' On Anniversary Of Tiananmen Square Crackdown (2021) (33)
15:32 Content Moderation Case Study: Twitter Removes 'Verified' Badge In Response To Policy Violations (2017) (8)
15:36 Content Moderation Case Study: Spam "Hacks" in Among Us (2020) (4)
15:37 Content Moderation Case Study: YouTube Deals With Disturbing Content Disguised As Videos For Kids (2017) (11)
15:48 Content Moderation Case Study: Twitter Temporarily Locks Account Of Indian Technology Minister For Copyright Violations (2021) (8)
15:45 Content Moderation Case Study: Spotify Comes Under Fire For Hosting Joe Rogan's Podcast (2020) (64)
15:48 Content Moderation Case Study: Twitter Experiences Problems Moderating Audio Tweets (2020) (6)
15:48 Content Moderation Case Study: Dealing With 'Cheap Fake' Modified Political Videos (2020) (9)
15:35 Content Moderation Case Study: Facebook Removes Image Of Two Men Kissing (2011) (13)
15:23 Content Moderation Case Study: Instagram Takes Down Instagram Account Of Book About Instagram (2020) (90)
15:49 Content Moderation Case Study: YouTube Relocates Video Accused Of Inflated Views (2014) (2)
15:34 Content Moderation Case Study: Pretty Much Every Platform Overreacts To Content Removal Stimuli (2015) (23)
16:03 Content Moderation Case Study: Roblox Tries To Deal With Adult Content On A Platform Used By Many Kids (2020) (0)
15:43 Content Moderation Case Study: Twitter Suspends Users Who Tweet The Word 'Memphis' (2021) (10)
15:35 Content Moderation Case Study: Time Warner Cable Doesn't Want Anyone To See Critical Parody (2013) (14)
15:38 Content Moderation Case Studies: Twitter Clarifies Hacked Material Policy After Hunter Biden Controversy (2020) (9)
15:42 Content Moderation Case Study: Kik Tries To Get Abuse Under Control (2017) (1)
15:31 Content Moderation Case Study: Newsletter Platform Substack Lets Users Make Most Of The Moderation Calls (2020) (8)
15:40 Content Moderation Case Study: Knitting Community Ravelry Bans All Talk Supporting President Trump (2019) (29)
15:50 Content Moderation Case Study: YouTube's New Policy On Nazi Content Results In Removal Of Historical And Education Videos (2019) (5)
15:36 Content Moderation Case Study: Google Removes Popular App That Removed Chinese Apps From Users' Phones (2020) (28)
15:42 Content Moderation Case Studies: How To Moderate World Leaders Justifying Violence (2020) (5)
15:47 Content Moderation Case Study: Apple Blocks WordPress Updates In Dispute Over Non-Existent In-app Purchase (2020) (18)
15:47 Content Moderation Case Study: Google Refuses To Honor Questionable Requests For Removal Of 'Defamatory' Content (2019) (25)
More arrow