YouTube is adding more human moderators and increasing its machine learning in an attempt to curb its child exploitation problem, the company's CEO Susan Wojcicki said. From a report: The company plans to increase its content moderation workforce to more than 10,000 employees in 2018 in order to help screen videos and train the platform's machine learning algorithms to spot and remove problematic children's content. Sources familiar with YouTube's workforce numbers say this represents a 25% increase from where the company is today. In the last two weeks, YouTube has removed hundreds of thousands of videos featuring children in disturbing and possibly exploitative situations, including being duct-taped to walls, mock-abducted, and even forced into washing machines. The company said it will employ the same approach it used this summer as it worked to eradicate violent extremist content from the platform.