{"id":3694,"date":"2021-03-18T23:33:50","date_gmt":"2021-03-18T22:33:50","guid":{"rendered":"https:\/\/technophilosoph.com\/2021\/04\/07\/gender-stereotyping-in-google-translate\/"},"modified":"2021-04-07T21:56:28","modified_gmt":"2021-04-07T20:56:28","slug":"gender-stereotyping-in-google-translate","status":"publish","type":"post","link":"https:\/\/technophilosoph.com\/en\/2021\/03\/18\/gender-stereotyping-in-google-translate\/","title":{"rendered":"Gender Stereotyping in Google Translate"},"content":{"rendered":"\n<p class=\"wp-block-paragraph\">Data bias caused by poorly curated data is increasingly emerging as an important issue in the development of artificial intelligence. To this end, it is important to know that today&#8217;s AI is fed millions and millions of data sets through machine learning. Where this data comes from, and who selects and compiles it, is critical. If such a data set, which is supposed to recognize faces, has predominantly faces with light skin color or more men than women, then such faces are recognized well, but the AI then makes more errors with women or dark-skinned people, or even in combination with the face recognition of dark-skinned women.<\/p>\n\n\n\n<p class=\"wp-block-paragraph\">While this may not be so tragic in the case of facial recognition, in other systems it can be a matter of life and death, imprisonment or freedom, or the granting or denial of credit. Another pitfall is that human stereotypes are embedded and deepened in the system. One such case was discovered by <a href=\"https:\/\/www.linkedin.com\/in\/annakholina\/\" target=\"_blank\" rel=\"noreferrer noopener\">Anna Kholina<\/a>, a Russian living in Helsinki. She had <a href=\"https:\/\/translate.google.com\/\" target=\"_blank\" rel=\"noreferrer noopener\">Google Translate<\/a>, an AI-based online translation software, translate some sentences from gender-neutral Finnish into English. The result reflected gender stereotypes. Sentences that the system associated with typically female activities were translated with the female case, and those with typically male activities were translated with the male case.<\/p>\n\n\n\n<figure class=\"wp-block-image size-large\"><img data-recalc-dims=\"1\" loading=\"lazy\" decoding=\"async\" width=\"676\" height=\"552\" src=\"https:\/\/i0.wp.com\/technophilosoph.com\/wp-content\/uploads\/2021\/03\/gt-f-e.jpg?resize=676%2C552&#038;ssl=1\" alt=\"\" class=\"wp-image-3630\"\/><\/figure>\n\n\n\n<p class=\"wp-block-paragraph\">Such translations pose a problem when it comes to eradicating gender stereotypes. The fact that women are allowed to vote, study, and take up professions, and even roles that were only open to men until recently can now be exercised as a matter of course, is not self-evident and also has a lot to do with language and pointing out opportunities.<\/p>\n\n\n\n<!--more-->\n\n\n\n<p class=\"wp-block-paragraph\">How it is that Google Translate spreads such stereotypes is quickly explained: the underlying texts, written by humans, until not so long ago perpetuated precisely these traditional gender roles. The woman stands at the hearth, the man goes to work. The woman takes care of children, the man plays sports. The AI learns from these sample texts that children and hearth are more often associated with woman (and her) than child and hearth with man. The system then selects the most likely combinations from these correlations. Conversely, negative stereotypes are also reinforced, as the example with junkie, stupid, and alcoholic shows. These characteristics translated as male.<\/p>\n\n\n\n<figure class=\"wp-block-image size-large\"><img data-recalc-dims=\"1\" loading=\"lazy\" decoding=\"async\" width=\"720\" height=\"164\" src=\"https:\/\/i0.wp.com\/technophilosoph.com\/wp-content\/uploads\/2021\/03\/gt-f-e-junkie.png?resize=720%2C164&#038;ssl=1\" alt=\"\" class=\"wp-image-3638\"\/><\/figure>\n\n\n\n<p class=\"wp-block-paragraph\">This is precisely why data selection is so important in today&#8217;s AI systems. In German, by the way, the result does not look much different:<\/p>\n\n\n\n<figure class=\"wp-block-image size-large\"><img data-recalc-dims=\"1\" loading=\"lazy\" decoding=\"async\" width=\"688\" height=\"600\" src=\"https:\/\/i0.wp.com\/technophilosoph.com\/wp-content\/uploads\/2021\/03\/gt-f-d.png?resize=688%2C600&#038;ssl=1\" alt=\"\" class=\"wp-image-3631\"\/><figcaption>With the German AI translation tool <a href=\"https:\/\/www.deepl.com\/translator\" target=\"_blank\" rel=\"noreferrer noopener\">DeepL<\/a> the result is a bit better, but not perfect either:<\/figcaption><\/figure>\n\n\n\n<figure class=\"wp-block-image size-large\"><img data-recalc-dims=\"1\" loading=\"lazy\" decoding=\"async\" width=\"574\" height=\"744\" src=\"https:\/\/i0.wp.com\/technophilosoph.com\/wp-content\/uploads\/2021\/03\/deepl-f-d.png?resize=574%2C744&#038;ssl=1\" alt=\"\" class=\"wp-image-3633\"\/><\/figure>\n\n\n\n<p class=\"wp-block-paragraph\">However, a difference was discovered by users who did not use the English-language user interface of Google Translate, but the Finnish one. In the latter, the translation tool offered both male and female variants.<\/p>\n\n\n\n<figure class=\"wp-block-gallery columns-2 is-cropped wp-block-gallery-1 is-layout-flex wp-block-gallery-is-layout-flex\"><ul class=\"blocks-gallery-grid\"><li class=\"blocks-gallery-item\"><figure><img data-recalc-dims=\"1\" loading=\"lazy\" decoding=\"async\" width=\"720\" height=\"501\" src=\"https:\/\/i0.wp.com\/technophilosoph.com\/wp-content\/uploads\/2021\/03\/gt-f-e-eui.png?resize=720%2C501&#038;ssl=1\" alt=\"\" data-id=\"3640\" data-link=\"https:\/\/technophilosoph.com\/gt-f-e-eui\/\" class=\"wp-image-3640\"\/><\/figure><\/li><li class=\"blocks-gallery-item\"><figure><img data-recalc-dims=\"1\" loading=\"lazy\" decoding=\"async\" width=\"720\" height=\"327\" src=\"https:\/\/i0.wp.com\/technophilosoph.com\/wp-content\/uploads\/2021\/03\/gt-f-e-fui.png?resize=720%2C327&#038;ssl=1\" alt=\"\" data-id=\"3641\" data-link=\"https:\/\/technophilosoph.com\/gt-f-e-fui\/\" class=\"wp-image-3641\"\/><\/figure><\/li><\/ul><\/figure>\n\n\n\n<p class=\"wp-block-paragraph\">By the way, the entire discussion is on <a href=\"https:\/\/www.linkedin.com\/posts\/activity-6774971399715987456-xkkp\" target=\"_blank\" rel=\"noreferrer noopener\">LinkedIn<\/a>.<\/p>\n\n\n\n<p class=\"wp-block-paragraph\">If you want to try it yourself, here are the Finnish phrases.<\/p>\n\n\n\n<p class=\"wp-block-paragraph\">h\u00e4n on kaunis<br>h\u00e4n ajaa autoa<br>h\u00e4n lukee<br>h\u00e4n harjaa hiuksia<br>h\u00e4n laskee<br>h\u00e4n tanssii<br>h\u00e4n urheilee<br>h\u00e4n puhuu<br>h\u00e4n sijoitaa<br>h\u00e4n k\u00e4y t\u00f6issa<br>h\u00e4n hoita lasta<br>h\u00e4n pesee pyykki\u00e4<br>h\u00e4n tiennaa rahaa<br>h\u00e4n on tyhm\u00e4<br>h\u00e4n on narkkari<\/p>\n","protected":false},"excerpt":{"rendered":"<p>Data bias caused by poorly curated data is increasingly emerging as an important issue in the development of artificial intelligence. To this end, it is important to know that today&#8217;s AI is fed millions and millions of data sets through &hellip; <a href=\"https:\/\/technophilosoph.com\/en\/2021\/03\/18\/gender-stereotyping-in-google-translate\/\" class=\"more-link\">Continue reading <span class=\"screen-reader-text\">Gender Stereotyping in Google Translate<\/span><\/a><\/p>\n","protected":false},"author":8770504,"featured_media":3701,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"_coblocks_attr":"","_coblocks_dimensions":"","_coblocks_responsive_height":"","_coblocks_accordion_ie_support":"","_crdt_document":"","advanced_seo_description":"","jetpack_seo_html_title":"","jetpack_seo_noindex":false,"_jetpack_newsletter_access":"","_jetpack_dont_email_post_to_subs":false,"_jetpack_newsletter_tier_id":0,"_jetpack_memberships_contains_paywalled_content":false,"_jetpack_memberships_contains_paid_content":false,"footnotes":"","jetpack_publicize_message":"","jetpack_publicize_feature_enabled":true,"jetpack_social_post_already_shared":true,"jetpack_social_options":{"image_generator_settings":{"template":"highway","default_image_id":0,"font":"","enabled":false},"version":2},"_wpas_customize_per_network":false,"jetpack_post_was_ever_published":false},"categories":[694240470,694240471,694240470,694240471],"tags":[],"class_list":{"0":"post-3694","1":"post","2":"type-post","3":"status-publish","4":"format-standard","5":"has-post-thumbnail","6":"hentry","7":"category-artificial-intelligence","8":"category-bias","12":"fallback-thumbnail"},"jetpack_publicize_connections":[],"jetpack_featured_media_url":"https:\/\/i0.wp.com\/technophilosoph.com\/wp-content\/uploads\/2021\/04\/Google-Translate-EN.jpeg?fit=1200%2C900&ssl=1","jetpack_likes_enabled":true,"jetpack_sharing_enabled":true,"jetpack_shortlink":"https:\/\/wp.me\/p72hdn-XA","amp_enabled":true,"_links":{"self":[{"href":"https:\/\/technophilosoph.com\/en\/wp-json\/wp\/v2\/posts\/3694","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/technophilosoph.com\/en\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/technophilosoph.com\/en\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/technophilosoph.com\/en\/wp-json\/wp\/v2\/users\/8770504"}],"replies":[{"embeddable":true,"href":"https:\/\/technophilosoph.com\/en\/wp-json\/wp\/v2\/comments?post=3694"}],"version-history":[{"count":2,"href":"https:\/\/technophilosoph.com\/en\/wp-json\/wp\/v2\/posts\/3694\/revisions"}],"predecessor-version":[{"id":3704,"href":"https:\/\/technophilosoph.com\/en\/wp-json\/wp\/v2\/posts\/3694\/revisions\/3704"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/technophilosoph.com\/en\/wp-json\/wp\/v2\/media\/3701"}],"wp:attachment":[{"href":"https:\/\/technophilosoph.com\/en\/wp-json\/wp\/v2\/media?parent=3694"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/technophilosoph.com\/en\/wp-json\/wp\/v2\/categories?post=3694"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/technophilosoph.com\/en\/wp-json\/wp\/v2\/tags?post=3694"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}