行业英语 学英语,练听力,上听力课堂! 注册 登录
> 行业英语 > 金融英语 > 金融时报原文阅读 >  第276篇

向机器人求个职

所属教程:金融时报原文阅读

浏览:

2020年06月29日

手机版
扫描二维码方便学习和分享

向机器人求个职

机器人并不只抢走人类的工作,它们也开始招聘人类员工了,因为它们可以快速筛选应聘者,但这很危险。

测试中可能遇到的词汇和知识:

the air is thick with空气里弥漫着

seductive有魅力的;性感的[sɪ'dʌktɪv]

inherently内在地;固有地;天性地[ɪnˈhɪərəntlɪ]

bias偏见;偏爱;斜纹;乖离率['baɪəs]

ethnicity种族划分[eθ'nɪsɪtɪ]

proxy代理人;委托书 ['prɒksɪ]

scenario剧本;设想[sɪ'nɑːrɪəʊ]

murky黑暗的;朦胧的;阴郁的['mɜːkɪ]

The risks of relying on robots for fairer staff recruitment(630 words)

By Sarah O’Connor

Robots are not just taking people’s jobs away,they are beginning to hand them out,too. Go to any recruitment industry event and you will find the air is thick with terms like“machine learning”,“big data”and“predictive analytics”.

The argument for using these tools in recruitment is simple. Robo-recruiters can sift through thousands of job candidates far more efficiently than humans. They can also do it more fairly. Since they do not harbour conscious or unconscious human biases,they will recruit a more diverse and meritocratic workforce.

This is a seductive idea but it is also dangerous. Algorithms are not inherently neutral just because they see the world in zeros and ones.

For a start,any machine learning algorithm is only as good as the training data from which it learns. Take the PhD thesis of academic researcher Colin Lee,released to the press this year. He analysed data on the success or failure of 441,769 job applications and built a model that could predict with 70 to 80 per cent accuracy which candidates would be invited to interview. The press release plugged this algorithm as a potential tool to screen a large number of CVs while avoiding“human error and unconscious bias”.

But a model like this would absorb any human biases at work in the original recruitment decisions. For example,the research found that age was the biggest predictor of being invited to interview,with the youngest and the oldest applicants least likely to be successful. You might think it fair enough that inexperienced youngsters do badly,but the routine rejection of older candidates seems like something to investigate rather than codify and perpetuate.

Mr Lee acknowledges these problems and suggests it would be better to strip the CVs of attributes such as gender,age and ethnicity before using them. Even then,algorithms can wind up discriminating. In a paper published this year,academics Solon Barocas and Andrew Selbst use the example of an employer who wants to select those candidates most likely to stay for the long term. If the historical data show women tend to stay in jobs for a significantly shorter time than men(possibly because they leave when they have children),the algorithm will probably discriminate against them on the basis of attributes that are a reliable proxy for gender.

Or how about the distance a candidate lives from the office? That might well be a good predictor of attendance or longevity at the company; but it could also inadvertently discriminate against some groups,since neighbourhoods can have different ethnic or age profiles.

These scenarios raise the tricky question of whether it is wrong to discriminate even when it is rational and unintended. This is murky legal territory. In the US,the doctrine of“disparate impact”outlaws ostensibly neutral employment practices that disproportionately harm“protected classes”,even if the employer does not intend to discriminate. But employers can successfully defend themselves if they can prove there is a strong business case for what they are doing. If the intention of the algorithm is simply to recruit the best people for the job,that may be a good enough defence.

Still,it is clear that employers who want a more diverse workforce cannot assume that all they need to do is turn over recruitment to a computer. If that is what they want,they will need to use data more imaginatively.

Instead of taking their own company culture as a given and looking for the candidates statistically most likely to prosper within it,for example,they could seek out data about where(and in which circumstances) a more diverse set of workers thrive.

Machine learning will not propel your workforce into the future if the only thing it learns from is your past.

1.What is not the reason of using these robots in recruitment?

A. sift job candidates more efficiently

B. take a more procedural approach to save time

C. sift job candidates more fairly

D. recruit a more diverse and meritocratic workforce

答案(1)

2.Which one is not right about relying on robots for fairer staff recruitment as mentioned?

A. algorithms are inherently neutral

B. it is seductive but dangerous

C. robots see the world in zeros and ones

D. machine learning algorithm is only as good as the training data from which it learns

答案(2)

3.What was the biggest predictor of being invited to interview of Colin Lee’s research?

A. gender

B. ethnicity

C. age

D. education

答案(3)

4.What should employers do if they want a more diverse workforce by computer recruitment?

A. use data more accurately

B. use data more imaginatively

C. gather more data

D. strip the CVs of attributes such as gender,age and ethnicity

答案(4)

(1) 答案:B.take a more procedural approach to save time

解释:机器人招聘者可以快速筛选数以千计的应聘者,效率远高于人类。它们还能做到更加公平。因为它们不会像人类那样带着有意或无意的偏见,它们会招聘到一批更多元化和择优录用的员工。

(2) 答案:A.algorithms are inherently neutral

解释:这是个很诱人的想法,但也是危险的。算法的中立并非是其固有,而是因为它们看到的世界只是“0”和“1”。任何机器学习的算法,并不会比它所学习的训练数据更好。

(3) 答案:C.age

解释:研究发现,年龄因素可以在最大程度上预测该应聘者是否会被邀请面试,最年轻和最年长的应聘者最不可能成功。

(4) 答案:B.use data more imaginatively

解释:那些希望把招聘交给电脑去做,又要拥有更多元化的员工队伍的雇主,应该把数据运用得更富想象力一些。

用户搜索

疯狂英语 英语语法 新概念英语 走遍美国 四级听力 英语音标 英语入门 发音 美语 四级 新东方 七年级 赖世雄 zero是什么意思吉林市大街小区英语学习交流群

网站推荐

英语翻译英语应急口语8000句听歌学英语英语学习方法

  • 频道推荐
  • |
  • 全站推荐
  • 推荐下载
  • 网站推荐