{"componentChunkName":"component---src-templates-blog-post-js","path":"/blog/demographic-skews-in-training-data-create-algorithmic-errors/","result":{"data":{"site":{"siteMetadata":{"title":"No Frills News"}},"contentfulNfnPost":{"postTitle":"Demographic skews in training data create algorithmic errors","slug":"demographic-skews-in-training-data-create-algorithmic-errors","createdLocal":"2021-06-04 14:31:28.343188","publishDate":"2021-06-05 00:00:00","feedName":"Image Recognition","sourceUrl":{"sourceUrl":"https://www.economist.com/graphic-detail/2021/06/05/demographic-skews-in-training-data-create-algorithmic-errors"},"postSummary":{"childMarkdownRemark":{"html":"<p>Every good model relies on training data that reflect what it seeks to predict.\nFor uses like facial recognition, models need enough cases from each demographic group to learn how to identify members accurately.\nAnd when making forecasts, like trying to predict successful hires from recorded job interviews, the proportions of each group in training data should resemble those in the population.\nMany businesses compile private training data.\nOne study of three programs that identify sex in photos found far more errors for dark-skinned women than for light-skinned men.</p>"}}}},"pageContext":{"slug":"demographic-skews-in-training-data-create-algorithmic-errors"}}}