zlacker

[parent] [thread] 1 comments
1. gt_+(OP)[view] [source] 2017-12-17 22:07:38
??? I just read the article... not sure what you're point is. My point is that the author "encoded and even amplified" bias that the data did not have. Most of the article concerns being aware of inherent bias in data. My point concerns bias that the author has ascribed to the data.
replies(1): >>nl+x
2. nl+x[view] [source] 2017-12-17 22:15:31
>>gt_+(OP)
Gender bias in Word2Vec data is a well known problem. The article provides references. It's an unsupervised algorithm, and the Google pretrained vectors are trained from news coverage, so not really a researcher issue (insofar as they select an unusual datasets or something).

Edit: to clarify, your claim is that Word2Vec data isn't biased even though there is a link right there showing how it is? Why do you think that?

If you use that data in a system then you reinforce that bias.

[go to top]