10.5555/3157382.3157584guideproceedingsArticle/Chapter ViewAbstractPublication PagesnipsConference Proceedingsconference-collections
Article
Free Access

Man is to computer programmer as woman is to homemaker? debiasing word embeddings

Online:05 December 2016Publication History

ABSTRACT

The blind application of machine learning runs the risk of amplifying biases present in data. Such a danger is facing us with word embedding, a popular framework to represent text data as vectors which has been used in many machine learning and natural language processing tasks. We show that even word embeddings trained on Google News articles exhibit female/male gender stereotypes to a disturbing extent. This raises concerns because their widespread use, as we describe, often tends to amplify these biases. Geometrically, gender bias is first shown to be captured by a direction in the word embedding. Second, gender neutral words are shown to be linearly separable from gender definition words in the word embedding. Using these properties, we provide a methodology for modifying an embedding to remove gender stereotypes, such as the association between the words receptionist and female, while maintaining desired associations such as between the words queen and female. Using crowd-worker evaluation as well as standard benchmarks, we empirically demonstrate that our algorithms significantly reduce gender bias in embeddings while preserving the its useful properties such as the ability to cluster related concepts and to solve analogy tasks. The resulting embeddings can be used in applications without amplifying gender bias.

References

  1. J. Angwin, J. Larson, S. Mattu, and L. Kirchner. Machine bias: There's software used across the country to predict future criminals. and it's biased against blacks., 2016.Google ScholarGoogle Scholar
  2. S. Barocas and A. D. Selbst. Big data's disparate impact. Available at SSRN 2477899, 2014.Google ScholarGoogle Scholar
  3. E. Beigman and B. B. Klebanov. Learning with annotation noise. In ACL, 2009. Google ScholarGoogle ScholarDigital LibraryDigital Library
  4. A. Datta, M. C. Tschantz, and A. Datta. Automated experiments on ad privacy settings. Proceedings on Privacy Enhancing Technologies, 2015.Google ScholarGoogle ScholarCross RefCross Ref
  5. C. Dwork, M. Hardt, T. Pitassi, O. Reingold, and R. Zemel. Fairness through awareness. In Innovations in Theoretical Computer Science Conference, 2012. Google ScholarGoogle ScholarDigital LibraryDigital Library
  6. J. Eisenstein, B. O'Connor, N. A. Smith, and E. P. Xing. Diffusion of lexical change in social media. PLoS ONE, pages 1-13, 2014.Google ScholarGoogle ScholarCross RefCross Ref
  7. M. Faruqui, J. Dodge, S. K. Jauhar, C. Dyer, E. Hovy, and N. A. Smith. Retrofitting word vectors to semantic lexicons. In NAACL, 2015.Google ScholarGoogle ScholarCross RefCross Ref
  8. M. Feldman, S. A. Friedler, J. Moeller, C. Scheidegger, and S. Venkatasubramanian. Certifying and removing disparate impact. In KDD, 2015. Google ScholarGoogle ScholarDigital LibraryDigital Library
  9. C. Fellbaum, editor. Wordnet: An Electronic Lexical Database. The MIT Press, Cambridge, MA, 1998.Google ScholarGoogle Scholar
  10. L. Finkelstein, E. Gabrilovich, Y. Matias, E. Rivlin, Z. Solan, G. Wolfman, and E. Ruppin. Placing search in context: The concept revisited. In WWW. ACM, 2001. Google ScholarGoogle ScholarDigital LibraryDigital Library
  11. A. G. Greenwald, D. E. McGhee, and J. L. Schwartz. Measuring individual differences in implicit cognition: the implicit association test. Journal of personality and social psychology, 74(6):1464, 1998.Google ScholarGoogle Scholar
  12. C. Hansen, M. Tosik, G. Goossen, C. Li, L. Bayeva, F. Berbain, and M. Rotaru. How to get the best word vectors for resume parsing. In SNN Adaptive Intelligence / Symposium: Machine Learning 2015, Nijmegen.Google ScholarGoogle Scholar
  13. J. Holmes and M. Meyerhoff. The handbook of language and gender, volume 25. John Wiley & Sons, 2008.Google ScholarGoogle Scholar
  14. O. İrsoy and C. Cardie. Deep recursive neural networks for compositionality in language. In NIPS. 2014. Google ScholarGoogle ScholarDigital LibraryDigital Library
  15. R. Jakobson, L. R. Waugh, and M. Monville-Burston. On language. Harvard Univ Pr, 1990.Google ScholarGoogle Scholar
  16. M. Kay, C. Matuszek, and S. A. Munson. Unequal representation and gender stereotypes in image search results for occupations. In Human Factors in Computing Systems. ACM, 2015. Google ScholarGoogle ScholarDigital LibraryDigital Library
  17. T. Lei, H. Joshi, R. Barzilay, T. Jaakkola, A. M. Katerina Tymoshenko, and L. Marquez. Semi-supervised question retrieval with gated convolutions. In NAACL. 2016.Google ScholarGoogle Scholar
  18. O. Levy and Y. Goldberg. Linguistic regularities in sparse and explicit word representations. In CoNLL, 2014.Google ScholarGoogle ScholarCross RefCross Ref
  19. T. Mikolov, K. Chen, G. Corrado, and J. Dean. Efficient estimation of word representations in vector space. In ICLR, 2013.Google ScholarGoogle Scholar
  20. T. Mikolov, I. Sutskever, K. Chen, G. S. Corrado, and J. Dean. Distributed representations of words and phrases and their compositionality. In NIPS. Google ScholarGoogle ScholarDigital LibraryDigital Library
  21. T. Mikolov, W.-t. Yih, and G. Zweig. Linguistic regularities in continuous space word representations. In HLT-NAACL, pages 746-751, 2013.Google ScholarGoogle Scholar
  22. E. Nalisnick, B. Mitra, N. Craswell, and R. Caruana. Improving document ranking with dual word embeddings. In www, April 2016. Google ScholarGoogle ScholarDigital LibraryDigital Library
  23. B. A. Nosek, M. Banaji, and A. G. Greenwald. Harvesting implicit group attitudes and beliefs from a demonstration web site. Group Dynamics: Theory, Research, and Practice, 6(1):101, 2002.Google ScholarGoogle Scholar
  24. D. Pedreshi, S. Ruggieri, and F. Turini. Discrimination-aware data mining. In SIGKDD, pages 560-568. ACM, 2008. Google ScholarGoogle ScholarDigital LibraryDigital Library
  25. J. Pennington, R. Socher, and C. D. Manning. Glove: Global vectors for word representation. In EMNLP, 2014.Google ScholarGoogle ScholarCross RefCross Ref
  26. K. Ross and C. Carter. Women and news: A long and winding road. Media, Culture & Society, 33(8):1148-1165, 2011.Google ScholarGoogle Scholar
  27. H. Rubenstein and J. B. Goodenough. Contextual correlates of synonymy. Communications of the ACM, 8(10):627-633, 1965. Google ScholarGoogle ScholarDigital LibraryDigital Library
  28. E. Sapir. Selected writings of Edward Sapir in language, culture and personality, volume 342. Univ of California Press, 1985.Google ScholarGoogle Scholar
  29. B. Schmidt. Rejecting the gender binary: a vector-space operation. http://bookworm.benschmidt.org/posts/2015-10-30-rejecting-the-gender-binary.html, 2015.Google ScholarGoogle Scholar
  30. J. P. Stanley. Paradigmatic woman: The prostitute. Papers in language variation, pages 303-321, 1977.Google ScholarGoogle Scholar
  31. L. Sweeney. Discrimination in online ad delivery. Queue, 11(3):10, 2013. Google ScholarGoogle ScholarDigital LibraryDigital Library
  32. A. Torralba and A. Efros. Unbiased look at dataset bias. In CVPR, 2012. Google ScholarGoogle ScholarDigital LibraryDigital Library
  33. P. D. Turney. Domain and function: A dual-space model of semantic relations and compositions. Journal of Artificial Intelligence Research, pages 533-585, 2012. Google ScholarGoogle ScholarDigital LibraryDigital Library
  34. C. Wagner, D. Garcia, M. Jadidi, and M. Strohmaier. It's a man's wikipedia? assessing gender inequality in an online encyclopedia. In Ninth International AAAI Conference on Web and Social Media, 2015.Google ScholarGoogle Scholar
  35. D. Yogatama, M. Faruqui, C. Dyer, and N. A. Smith. Learning word representations with hierarchical sparse coding. In ICML, 2015. Google ScholarGoogle ScholarDigital LibraryDigital Library
  36. I. Zliobaite. A survey on measuring indirect discrimination in machine learning. arXiv preprint arXiv:1511.00148, 2015.Google ScholarGoogle Scholar

Index Terms

(auto-classified)
  1. Man is to computer programmer as woman is to homemaker? debiasing word embeddings

    Comments

    Login options

    Check if you have access through your login credentials or your institution to get full access on this article.

    Sign in
    • Published in

      Guide Proceedings cover image
      NIPS'16: Proceedings of the 30th International Conference on Neural Information Processing Systems
      December 2016
      5100 pages

      Publisher

      Curran Associates Inc.

      Red Hook, NY, United States

      Publication History

      • Online: 5 December 2016
      • Published: 5 December 2016

      Qualifiers

      • Article

    PDF Format

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader
    About Cookies On This Site

    We use cookies to ensure that we give you the best experience on our website.

    Learn more

    Got it!