伦理网站_爱的伦理

  • Post author:
  • Post category:其他


伦理网站


This is part 1/3 in my series on The Application of AI.


这是我关于AI应用的系列文章的1/3。

Most people view

Artificial [Narrow] Intelligence

(AI) in an

idyllic

way, a tool that we can use to automate tasks, beat us in games, and

improve our lives in general!

大多数人以一种

田园诗般的

方式看待

人工智能[

AI

]

,这是我们可以用来自动执行任务,在游戏中击败我们并

改善我们的生活的工具!

Image for post


Andy Kelly

on

安迪·凯利

(

Unsplash


Uny Splash)摄

While this is largely true, there are also

many flaws and dangers of AI

, whether put in

intentionally or unintentionally

. Even today, there are countless examples of AI being misused to disrupt and damage the lives of many, so

why and how is this happening?

In the future, when AI is more widespread and common, what

potential

does it have to really

devastate our society?

In this article, I will be diving into the subject of

The


Ethics of AI

.

尽管这在很大程度上是正确的,但无论是

有意还是无意地

放置

AI

,也存在

许多缺陷和危险

。 即使在今天,仍然有无数的例子被滥用以破坏和破坏许多人的生命,

为什么会这样以及如何发生呢?

将来,当AI变得更加普遍和普遍时,它真正具有

摧毁我们社会的


潜力

是什么



在这篇文章中,我将深入探讨

人工智能



伦理

问题。

AI算法中的道德缺陷示例

(

Examples of Ethical Flaws in AI Algorithms

)

There are many unintentional flaws in AI that create ethical problems. These flaws could lie anywhere from the bias in the training data to the creation of feedback loops. Let’s look at some examples of how these flaws can affect people in the real world.

AI中有许多无意中的缺陷,这些缺陷会造成道德问题。 这些缺陷可能存在于从训练数据的偏差到创建反馈循环的任何地方。 让我们看一下这些缺陷如何影响现实世界中的人们的一些示例。

面部识别训练数据中的偏见

(

Bias in the Training Data of Facial Recognition

)


Imagine

one day you were coming home from work (or school), and the police show up. They show you an arrest warrant, handcuff you, and take you to the detention center. This is all happening while you have no idea what you have done, and your family and neighbors are watching you. When you arrive at the station they pat you down, fingerprint you, take a mugshot, then throw you in a dirty cell overnight. You sit there waiting, wondering why you were arrested. You’re a good person, you haven’t done anything wrong, and

you just want to see your family

.


想象

有一天,您下班(或上学)回家时,警察出现了。 他们会向您显示逮捕令,将您铐上手铐,然后带您到拘留所。 这一切都是在您不知道自己做了什么以及您的家人和邻居在看着您的时候发生的。 当您到达车站时,他们会拍打您,为您打上指纹,进行面部照像,然后将您放在肮脏的牢房中过夜。 您坐在那里等待,想知道您为什么被捕。 你是一个好人,没有做错任何事,

只想见你的家人

Image for post


Detroit Free Press


底特律自由报的

罗伯特·威廉姆斯和他的家人

This is exactly

what happened to Robert Williams

, an innocent black man who was

wrongfully arrested

in Farmington Hills, Michigan. When the police were interrogating Williams they showed him a picture of a black man in a store, shoplifting valuable items. Williams denied he had been in the store and was not that man. The next photo was a close-up of the shoplifter, a photo that looked nothing like Williams. Williams had said:

这正是

罗伯特·威廉姆斯

(

Robert Williams)所发生的事情

,他是无辜的黑人,他在密歇根州的法明顿希尔斯被

错误逮捕

。 当警察审问威廉姆斯时,他们给他看了一张商店里的黑人男子的照片,他们在偷窃有价值的物品。 威廉姆斯否认他曾在商店里,不是那个男人。 下一张照片是入店行窃者的特写镜头,这张照片看上去完全不像威廉姆斯。 威廉姆斯曾说过:

“No, this is not me. You think all black men look alike?”

“不,这不是我。 你以为所有黑人看起来都一样吗?”

So how could the police be so wrong to arrest the wrong person? In this case, it was an algorithm’s fault. The police were using

AI to find suspects for crimes

, and obviously it was

faulty

. But how could an AI be wrong? Aren’t they supposed to be very accurate?

那么警察怎么可能错到逮捕错误的人呢? 在这种情况下,这是算法的错误。 警察正在使用

AI寻找犯罪嫌疑人

,显然这是

错误的

。 但是,人工智能怎么可能是错误的呢? 他们不应该很准确吗?

Yes, AI algorithms are usually very accurate, but only if you train them correctly. The main factor which led this algorithm to misidentify Williams was

bias in the training data

. While facial