正如物理学家、曼哈顿计划领导者奥本海默所说:“我们不仅是科学家,我们也是人。”,有人的地方就有江湖,科学界也难免。人们经常说“科学无国界,科学家有祖国“,即使不谈这些政治纠葛,科学家中还是有不同学术圈子的,每个人都有自己坚持的学术观点和主张,学术争论无时不在。一般来说,自由争论有利于学术进步,但争论也可能造成误会产生偏见,从而让某些科学家受害并影响科学的正常发展。今天将讲述的,是AI历史中的一段江湖故事……
As the physicist and leader of the Manhattan Project, Oppenheimer, said: "We are not only scientists, we are also human." Where there are people, there are conflicts, and the scientific community is no exception. People often say "Science knows no borders, but scientists have a homeland." Even without discussing these political entanglements, there are still different academic circles among scientists, each with their own academic views and propositions. Academic debates are constant. Generally speaking, free debate is conducive to academic progress, but debates can also lead to misunderstandings and biases, which may harm certain scientists and affect the normal development of science. Today, we will tell a story from the history of AI...
1958年7月,美国海军研究办公室公布了一项非凡的发明,
In July 1958, the U.S. Navy Research Office announced an extraordinary invention, claiming to demonstrate "the first machine capable of human thought," as shown in Figure 1.
演示者将一系列打孔卡经过一个电子设备,输入到一台重5吨,大小相当于一个房间的计算机(IBM704)中,经过50次试验后,计算机学会了区分左侧标记的卡片和右侧标记的卡片。
The demonstrator fed a series of punched cards through an electronic device into a computer (IBM704) weighing 5 tons and the size of a room. After 50 trials, the computer learned to distinguish between cards marked on the left and those marked on the right. In other words, this machine could learn to "classify," much like a child learns to distinguish between cats and dogs under parental guidance. Classification is an important function in artificial intelligence research.
图1:罗森布拉特和感知器Mark-1 Figure 1: Rosenblatt and the Perceptron Mark-1
The U.S. Navy demonstrated the "Perceptron." According to its creator, Dr. Frank Rosenblatt, it is an electronic device built on the principles of "neural networks" in biology, with learning capabilities. Rosenblatt detailed and expanded this method in his 1962 book "Principles of Neurodynamics: Perceptrons and the Theory of Brain Mechanisms." That year, Rosenblatt gained international recognition for the Perceptron. The New York Times called it a revolution, with the headline "New Navy Device Learns by Doing," and The New Yorker also praised this technological advancement.
当时的罗森布拉特是纽约布法罗康奈尔航空实验室的研究心理学家和项目工程师,展示感知机后的第二年,他成为康奈尔大学生物科学系神经生物学和行为学副教授。
At that time, Rosenblatt was a research psychologist and project engineer at the Cornell Aeronautical Laboratory in Buffalo, New York. The year after demonstrating the Perceptron, he became an associate professor of neurobiology and behavior in the Department of Biological Sciences at Cornell University.
罗森布拉特的感知机计划受到了McCulloch和Pitts的形式神经网络的启发,是专为图像识别而设计的一个简单的单层神经网络,并且添加了额外的机器学习机制。图2是感知机的逻辑示意图。感知机获得了美国海军的大量资助,最初是基于两年的软件研究而实现。之后,罗森布拉特构建并展示了感知器的唯一硬件版本:Mark-1,它由联结到神经元的400个光电细胞(20×20的感光单元矩阵)组成,可以将输入的光学信号(例如英文字符)转化为电信号,再通过物理电缆将其与字母分类的神经元层相连。Mark-1突触的权重函数用电位计编码,由电动马达实现学习过程中权重的改变。
Rosenblatt's Perceptron project was inspired by the formal neural networks of McCulloch and Pitts and was designed as a simple single-layer neural network specifically for image recognition, with additional machine learning mechanisms. Figure 2 is a logical schematic of the Perceptron. The Perceptron received substantial funding from the U.S. Navy and was initially implemented based on two years of software research. Subsequently, Rosenblatt built and demonstrated the only hardware version of the Perceptron: the Mark-1, which consisted of 400 photocells (a 20×20 matrix of photosensitive units) connected to neurons. It could convert input optical signals (such as English characters) into electrical signals, which were then connected to a layer of neurons for letter classification via physical cables. The weight functions of the Mark-1 synapses were encoded with potentiometers, and changes in weights during the learning process were achieved by electric motors.
可想而知,在当时的技术条件下,这台机器的具体实现还是非常困难的,这也是为什么引起轰动和广泛关注的原因。
It is conceivable that, given the technological conditions of the time, the specific implementation of this machine was very challenging, which is why it caused a sensation and attracted widespread attention.
图2:感知器设计概念图(1958年) Figure 2: Perceptron Design Concept Diagram (1958)
罗森布拉特对他的感知器寄予厚望,对人工智能神经网络的研究持十分乐观的态度,他乐观地认为突破即将到来。原来十分低调的科学家突然走红,他出席各种演讲、晚会,这点当然会引起当年人工智能大伽们的注意。
Rosenblatt had high hopes for his perceptron and was very optimistic about the research on artificial intelligence neural networks. He optimistically believed that a breakthrough was imminent. The previously low-profile scientist suddenly became popular, attending various lectures and parties, which naturally caught the attention of the AI bigwigs of the time.
罗森布拉特的工作引起了MIT闵斯基教授的关注。闵斯基在感知器问世的两年之前,即1956年,与麦卡锡等一起,发起召开了达特茅斯研讨会,确定了人工智能的名字,讨论了发展方向等等问题。基于他对神经网络的研究,使得他对罗森布拉特的主张深感怀疑。科学中的质疑是正常现象,因此,他们经常在学术会议上公开辩论感知机的可行性。在一次会议上,两人大吵一顿,矛盾彻底公开化。想必那几次争论是非常激烈的,因为据他们的同事和学生在后来的回忆中,有“在一旁看得目瞪口呆”、“被他们的争论吓了一跳”之类的话语。闵斯基直接对感知机存在的价值和前途发起进攻,指出它的实际价值非常有限,没有什么发展前途,绝不可能作为解决人工智能的问题的主要研究方法。
Rosenblatt's work caught the attention of MIT Professor Marvin Minsky. Two years before the perceptron was introduced, in 1956, Minsky, along with McCarthy and others, initiated the Dartmouth Conference, which established the name of artificial intelligence and discussed issues such as its development direction. Based on his research on neural networks, he was deeply skeptical of Rosenblatt's claims. Skepticism in science is normal, so they often publicly debated the feasibility of the perceptron at academic conferences. At one conference, the two had a heated argument, making their conflict completely public. Those debates must have been very intense, as colleagues and students later recalled being "dumbfounded" and "shocked by their arguments." Minsky directly attacked the value and future of the perceptron, pointing out that its practical value was very limited, with no development prospects, and it could never be the main research method for solving artificial intelligence problems.
“罗森布拉特认为他可以使计算机阅读并理解语言,而马文·闵斯基指出这不可能,因为感知机的功能太简单了”一位当年的研究生回忆说。
"Rosenblatt believed he could make computers read and understand language, while Marvin Minsky pointed out that this was impossible because the perceptron's functions were too simple," recalled a graduate student from that time.
后来,1969年,闵斯基和MIT的另一位数学教授西摩·佩珀特(Seymour Papert)出了一本名为《感知机》(Perceptrons)的学术著作[2],从理论上证明感知机的缺陷,书中还有对罗森布拉特个人的攻击言语:“罗森布拉特写的论文,大部分内容……毫无科学价值”。
Later, in 1969, Minsky and another MIT mathematics professor, Seymour Papert, published an academic book titled "Perceptrons," which theoretically demonstrated the flaws of the perceptron. The book also contained personal attacks on Rosenblatt, stating, "Most of the content in Rosenblatt's papers... has no scientific value."
在《感知机》这本书对罗森布拉特工作的强烈抨击下,本质上终结了感知机的命运。次年,闵斯基获得图灵奖,得到了计算机领域的最高荣誉。
Under the strong criticism of Rosenblatt's work in the book "Perceptrons," the fate of the perceptron was essentially sealed. The following year, Minsky received the Turing Award, the highest honor in the field of computer science.
闵斯基是当时业界的权威人物,这种对感知机直截了当的负面评价,对本性孤傲的罗森布拉特来说是致命的。一年多后,罗森布拉特在独自划船庆祝自己43岁生日那天溺水身亡,把他的名字,他的感知机,他的遗憾和梦想,都永远留在了人工智能的科学史上。
Minsky was an authoritative figure in the industry at the time, and such straightforward negative comments about the perceptron were fatal to the inherently proud Rosenblatt. More than a year later, Rosenblatt drowned while rowing alone to celebrate his 43rd birthday, leaving his name, his perceptron, his regrets, and dreams forever in the history of artificial intelligence.
《感知机》一书不仅打击了罗森布拉特,造成感知机的暂时失败,还几乎扼杀了当时神经网络方面的研究,也影响和带来了人工智能长达10年的第一次低谷期。
The book "Perceptrons" not only struck Rosenblatt, causing the temporary failure of the perceptron, but also almost killed research in neural networks at the time, leading to the first 10-year downturn in artificial intelligence.
2.符号主义和联结主义 2. Symbolism and Connectionism
其实,罗森布拉特[3]和闵斯基的出身和经历,有许多相同之处,他们年龄相仿,都是出生于纽约的犹太家庭,他们甚至曾经同时就读于同一所中学,是纽约布朗克斯科学高中的校友。不错!就是那所出了8名诺贝尔科学奖和一位诺贝尔经济奖得主,还有难以数计的各方名人的著名高中。如此名校的学兄学弟,却结怨于学海江湖!不由得使人脑海中划过那句名言:“相煎何太急?”
In fact, Rosenblatt[3] and Minsky had many similarities in their backgrounds and experiences. They were of similar age, both born into Jewish families in New York, and they even attended the same high school at the same time, being alumni of the Bronx High School of Science in New York. Yes! The very school that produced 8 Nobel Prize winners in science and one Nobel Prize winner in economics, along with countless famous figures from various fields. Yet, these alumni from such a prestigious school became adversaries in the academic world! It inevitably brings to mind the famous saying: "Why must brothers fight each other so urgently?"
不过,罗森布拉特去世后,闵斯基在《感知机》一书再版时,删除了原版中对罗森布拉特的个人攻击的句子,并手写了“纪念罗森布拉特”(In memory of Frank Rosenblatt)一语,多少表示了对这位早逝校友同行的哀悼。
However, after Rosenblatt's death, Minsky removed the sentences attacking Rosenblatt personally from the original edition of the book "Perceptrons" when it was reprinted, and handwrote the phrase "In memory of Frank Rosenblatt"(In memory of Frank Rosenblatt), somewhat expressing his condolences for his prematurely deceased fellow alumnus.
此外,两人的争论也代表了当年人工智能中符号主义与联结主义两种学术观点之争[4]。
In addition, the debate between the two also represented the academic dispute between symbolic and connectionist views in artificial intelligence at that time [4].
马文·闵斯基(Marvin Minsky,1927—2016)生于纽约市,他是深度学习的先行者。在哈佛大学读本科期间,他曾开发了早期的电子学习网络。在普林斯顿大学念研究生时,他建造了第一台神经网络学习机SNARC。他的博士论文以《神经-模拟强化系统的理论及其在大脑模型问题上的应用》为题,这实际上就是一篇关于神经网络的论文。所以,闵斯基研究生阶段的工作,奠定了人工神经网络的研究基础,应该属于联结主义的范畴。
Marvin Minsky (1927—2016) was born in New York City and was a pioneer of deep learning. During his undergraduate studies at Harvard University, he developed an early electronic learning network. While a graduate student at Princeton University, he built the first neural network learning machine, SNARC. His doctoral thesis was titled "Theory of Neural-Analog Reinforcement Systems and Its Application to the Brain Model Problem," which was essentially a paper on neural networks. Therefore, Minsky's work during his graduate studies laid the foundation for research in artificial neural networks and should be considered part of the connectionist domain.
1956年,他与约翰·麦卡锡、克劳德·香农等,一同发起了1956年的达特茅斯学院会议,创造了“人工智能”一词,是AI的开山鼻祖之一。达特茅斯学院会议也是符号主义的胜利,闵斯基和麦卡锡都被认为是AI符号主义的典型代表人物,当年他们的意图,是反对早期控制论的联结主义。认为人工智能的目的是通过程序在计算机中实现规则,用逻辑推理来对抗AI中的联结主义。从20世纪60年代中期到90年代初,符号主义方法盛行。
In 1956, he, along with John McCarthy, Claude Shannon, and others, initiated the Dartmouth Conference of 1956, coining the term "artificial intelligence," making him one of the pioneers of AI. The Dartmouth Conference was also a victory for symbolism, and Minsky and McCarthy were considered typical representatives of AI symbolism. Their intention at the time was to oppose the connectionism of early cybernetics, believing that the purpose of artificial intelligence was to implement rules in computers through programs and to use logical reasoning to counter connectionism in AI. From the mid-1960s to the early 1990s, symbolic methods prevailed.
可见,闵斯基后来转向了符号派,他也尽力淡化他与联结主义间的关系,也许这是他强烈抨击感知机的原因之一。
It is evident that Minsky later turned to the symbolic camp, and he also tried to downplay his relationship with connectionism, which may be one of the reasons for his strong criticism of perceptrons.
闵斯基自1958年起在麻省理工学院任教,担任电子工程与计算机科学的教授,直到他过世为止。
Minsky taught at the Massachusetts Institute of Technology from 1958, serving as a professor of electrical engineering and computer science until his death.
在麻省理工学院,他与约翰·麦卡锡共同创立了人工智能研究室(MIT计算机科学与人工智能实验室的前身)。他有数项发明,如1957年发表的共聚焦显微镜,1963年发表的头戴式显示器等。
At the Massachusetts Institute of Technology, he co-founded the Artificial Intelligence Laboratory (the predecessor of the MIT Computer Science and Artificial Intelligence Laboratory) with John McCarthy. He had several inventions, such as the confocal microscope published in 1957 and the head-mounted display published in 1963.
2016年1月24日,闵斯基因脑内出血病逝,享寿88岁。
On January 24, 2016, Minsky passed away due to a cerebral hemorrhage at the age of 88.
闵斯基的对手弗兰克·罗森布拉特(Frank Rosenblatt,1928-1971)比他小一岁,是一位心理学家。
Minsky's rival Frank Rosenblatt (1928-1971) was one year younger than him and was a psychologist.
罗森布拉特出生于纽约长岛的一个犹太家庭,1946年从布朗克斯科学高中毕业后,他进入康奈尔大学,并于1950年获得学士学位,1950年获得博士学位。随后,他前往纽约州布法罗的康奈尔航空实验室,先后担任研究心理学家、高级心理学家和认知系统部门负责人。这也是他进行感知器早期工作的地方。
Rosenblatt was born into a Jewish family on Long Island, New York. After graduating from Bronx High School of Science in 1946, he entered Cornell University, where he received a bachelor's degree in 1950 and a doctorate in 1950. He then went to the Cornell Aeronautical Laboratory in Buffalo, New York, where he served as a research psychologist, senior psychologist, and head of the cognitive systems department. This was also where he conducted early work on the perceptron.
罗森布拉特1966年加入康奈尔大学新成立的生物科学系神经生物学和行为科,担任副教授。他对通过注射脑提取物将学习行为从经过训练的老鼠转移到小老鼠身上产生了浓厚的兴趣,他在后来的几年中就这一主题发表了大量文章。
In 1966, Rosenblatt joined the newly established Department of Neurobiology and Behavior in the Division of Biological Sciences at Cornell University as an associate professor. He developed a strong interest in transferring learned behavior from trained rats to young rats through the injection of brain extracts, and he published numerous articles on this topic in the following years.
罗森布拉特也对天文学感兴趣,他花了3000美元买了一台望远镜,但太大了以至于没有地方放。因此,他在纽约布鲁克顿代尔附近买了一栋大房子,并邀请他的几名研究生住在那里。白天,团队在托伯莫里工作。晚上,他们在罗森布拉特的院子里作土木工作,建了一座天文台。
Rosenblatt was also interested in astronomy. He spent $3,000 on a telescope, but it was too large to fit anywhere. As a result, he bought a large house near Brocktondale, New York, and invited several of his graduate students to live there. During the day, the team worked in Tobermory. At night, they did civil work in Rosenblatt's yard and built an observatory.
罗森布拉特兴趣广泛多才多艺,白天在实验室里解剖蝙蝠,研究动物大脑的学习机理,夜晚在自家后山搭建的简易天文台上仰望天空,试图探索外星人奥秘。罗森布拉特的性格方面,害羞内向,并不张扬。
Rosenblatt had a wide range of interests and was versatile. During the day, he dissected bats in the laboratory to study the learning mechanisms of animal brains. At night, he gazed at the sky from a makeshift observatory he built on the hill behind his house, trying to explore the mysteries of aliens. In terms of personality, Rosenblatt was shy and introverted, not flamboyant.
感知器始终是罗森布拉特的热情所在。他最终没有熬过人工智能的冬天,于1971年43岁生日那天,自驾驶帆船溺水身亡。2004年,IEEE计算智能学会设立了罗森布拉特奖(IEEE Frank Rosenblatt Award),奖励在生物及语音启发计算领域做出卓越贡献的人,以纪念这位杰出的科学家。
The perceptron was always Rosenblatt's passion. He ultimately did not survive the AI winter and drowned on his 43rd birthday in 1971 while sailing. In 2004, the IEEE Computational Intelligence Society established the Rosenblatt Award (IEEE Frank Rosenblatt Award) to honor those who have made outstanding contributions in the field of biologically and linguistically inspired computing, in memory of this remarkable scientist.
图3:当年的感知机相关文章和书 Figure 3: Articles and books on the perceptron from that time
1956年的达特茅斯会议,启动了第一波人工智能浪潮,这一浪潮跨越至70年代初,代表符号主义的建模推理方法是其核心特征。这方面研究的主流由MIT的闵斯基、卡内基梅隆大学的西蒙和纽厄尔,以及斯坦福大学的麦卡锡组成。在当时,这个符号主义圈子里的高手们,基本建立了对人工智能问题的垄断,并获得了大部分资金和大型计算机系统的访问权。
The 1956 Dartmouth Conference initiated the first wave of artificial intelligence, which spanned until the early 1970s. The core feature of this wave was the symbolic modeling and reasoning methods. The mainstream research in this area was led by Minsky from MIT, Simon and Newell from Carnegie Mellon University, and McCarthy from Stanford University. At that time, the experts in this symbolic circle essentially monopolized the AI problem and gained access to most of the funding and large computer systems.
符号主义者的主要特征是他们不太重视机器智能与世界的联系,只在计算机内开辟独立的推理空间,将人工智能视为机器思维的科学,目标是赋予机器以逻辑和抽象的能力。
The main characteristic of symbolists is that they do not place much emphasis on the connection between machine intelligence and the world. They only create an independent reasoning space within the computer, viewing artificial intelligence as the science of machine thinking, with the goal of endowing machines with logical and abstract capabilities.
反之,罗森布拉特是心理学家,对人类生理学和心理行为等更感兴趣,因而趋向于联结主义。自然地,他热衷于对用神经网络的概念来模拟人脑神经传递机制,也由此而研究发明了感知机。
Conversely, Rosenblatt was a psychologist who was more interested in human physiology and psychological behavior, thus tending towards connectionism. Naturally, he was keen on using the concept of neural networks to simulate the neural transmission mechanisms of the human brain, which led him to research and invent the perceptron.
感知机在媒体上取得的成功,也激发起联结主义研究人员的热情。但闵斯基和帕佩特在1969年的书中宣称他们证明了神经网络无效的说法,又给这些科学家们浇了一盆冷水,使联结主义的热度骤减。尽管这本书的影响可能超出了闵斯基等的意图,但其后果是确定的:神经网络被放弃,其资金被大量削减。实际上,不仅联结主义衰退,针对符号主义的批评也越来越多,符号主义和联结主义项目都被冻结了,联邦对人工智能研究的资助枯竭。人工智能被当成仅仅是人工游戏,进入了它发展旅程中的第一个冬天。
The success of the perceptron in the media also inspired the enthusiasm of connectionist researchers. However, the claim by Minsky and Papert in their 1969 book that they had proven neural networks to be ineffective poured cold water on these scientists, causing the fervor for connectionism to plummet. Although the impact of this book may have exceeded Minsky and others' intentions, the consequences were certain: neural networks were abandoned, and their funding was significantly reduced. In fact, not only did connectionism decline, but criticism of symbolism also increased, leading to both symbolic and connectionist projects being frozen, and federal funding for artificial intelligence research dried up. Artificial intelligence was regarded merely as an artificial game, entering the first winter of its development journey.
3.感知器和神经网络 3. Perceptrons and Neural Networks
我们回到罗森布拉特的感知机[5]。它实际上是现代神经网络的雏形,有没有科学价值,已有如今AI的迅猛发展为证。当然,作为第一代的人工智能机器,感知器有这样那样的缺陷是难免的,而且当时的罗森布拉特,还没来得及把感知机的学习算法推广到多层神经网络。神经网络从简单到复杂多种多样,见图4。感知器只是一个最简单只有一层的神经网络(图4左),而现代神经网络却有数百万个(隐藏)层,图4右。
We return to Rosenblatt's perceptron. It is actually the prototype of modern neural networks, and its scientific value is evidenced by the rapid development of AI today. Of course, as the first generation of artificial intelligence machines, the perceptron inevitably had various flaws, and at that time, Rosenblatt had not yet had the opportunity to extend the learning algorithm of the perceptron to multilayer neural networks. Neural networks vary from simple to complex, as shown in Figure 4. The perceptron is just the simplest single-layer neural network (left of Figure 4), while modern neural networks have millions of (hidden) layers, as shown on the right of Figure 4.
图4:感知机和复杂神经网络 Figure 4: Perceptron and Complex Neural Networks
不过,闵斯基认为感知器的缺陷是致命的,因为它无法模拟“非线性可分”函数,他举了一个逻辑门的例子:异或门,即感知器不能区分异或门。以下对此作简单介绍。
However, Minsky believed that the flaw of the perceptron was fatal because it could not simulate "non-linearly separable" functions. He gave an example of a logic gate: the XOR gate, which the perceptron cannot distinguish. A brief introduction is given below.
感知器神经元的简单模型如图4左图所示:多个输入和一个输出。输出功能是求得输入向量与权向量的内积后,经一个激活函数得到一个标量结果。
The simple model of a perceptron neuron is shown in the left part of Figure 4: multiple inputs and one output. The output function is to obtain a scalar result by taking the inner product of the input vector and the weight vector, followed by an activation function.
神经网络为什么能分类呢?原因之一是归于激活函数的功劳。例如,最简单的激活函数是个阶梯函数,输出0或1,也就是说,这个函数实现了分类:将结果分成了两类。
Why can neural networks classify? One reason is attributed to the role of the activation function. For example, the simplest activation function is a step function, outputting 0 or 1. In other words, this function achieves classification: dividing the results into two categories.
至于何时输出0,何时输出1?就要根据输入的值来进行决策了。例如,可以问3个问题来判定是猫还是狗?耳朵朝上还是朝下?嘴巴是否凸出来?胡须长或短?最简单的决策方法就是:3个问题都输入yes的话,输出=猫,否则是“狗”。但激活函数可以从阶梯函数改变为平滑的函数,如图4左图中右下角的红线所示,这种函数便于在最优化时进行微分计算,而输出便被相应地理解为决策判断为是猫还是狗的概率了。
As for when to output 0 and when to output 1? It depends on the input values to make a decision. For example, you can ask three questions to determine if it's a cat or a dog: Are the ears up or down? Is the mouth protruding? Are the whiskers long or short? The simplest decision method is: if all three questions are answered yes, the output is a cat; otherwise, it's a "dog." However, the activation function can change from a step function to a smooth function, as shown by the red line in the lower right corner of the left part of Figure 4. This type of function facilitates differentiation during optimization, and the output is correspondingly understood as the probability of the decision being a cat or a dog.
神经网络又为何能具有学习功能呢?那是因为每个输入端都有一个权重值,这些参数是神经网络的核心。在训练过程中,网络调整这些权重,以使其在特定任务上的误差最小化。这个权重更新的过程也就是所谓的“机器学习”的过程。最小化可以使用各种优化算法,例如,感知器中用的是“梯度下降法”。
Why can neural networks have learning capabilities? This is because each input has a weight value, and these parameters are the core of the neural network. During training, the network adjusts these weights to minimize errors on specific tasks. This process of updating weights is what is known as "machine learning." Minimization can be achieved using various optimization algorithms, such as the "gradient descent method" used in perceptrons.
如图4左图上方的公式所示,为输出而计算的求和函数是n维空间中的一个超平面。因此,感知器这种神经网络“分类”的本质就是这个超平面将空间分成了两部分。如果对两个输入端的神经网络而言,就是用一条直线将平面分成了2个部分,如图5b所示的线性可分情况。
As shown in the formula at the top left of Figure 4, the summation function calculated for the output is a hyperplane in n-dimensional space. Therefore, the essence of the perceptron, a type of neural network "classification," is that this hyperplane divides the space into two parts. For a neural network with two inputs, it is like using a straight line to divide the plane into two parts, as shown in the linearly separable case in Figure 5b.
图5:感知机分类,线性可分和不可分 Figure 5: Perceptron classification, linearly separable and inseparable
然而,如果输入的样本是线性不可分的(图5b右边),感知机则无法模拟这种情况。这就是闵斯基指出的感知机的缺点。
However, if the input samples are linearly inseparable (right side of Figure 5b), the perceptron cannot model this situation. This is the drawback of the perceptron pointed out by Minsky.
图6显示的是几种基本逻辑门的情况,单层感知机可被用来区分其中的3种,逻辑与(AND)、逻辑与非(NAND)和逻辑或(OR),但是,无法模拟逻辑异或函数,因为它属于线性不可分。
Figure 6 shows the cases of several basic logic gates. A single-layer perceptron can be used to distinguish three of them: AND, NAND, and OR, but it cannot model the XOR function because it is linearly inseparable.
图6:逻辑门,前3种是线性可分的,XOR非线性可分
Figure 6: Logic gates, the first three types are linearly separable, XOR is non-linearly separable
要解决非线性可分问题,可考虑使用多层功能神经网络。输出层与输入层之间的一层神经元,被称为隐层,隐层和输出层神经元都是拥有激活函数的功能神经元。图7左图中,感知器的神经元,没有隐含层,决策计算只生成一条直线,无法区别异或问题。但如果增加一个带非线性激活函数的隐含层就可以了。隐层输出的激活函数的非线性也有助于解决非线性可分问题。多一个隐层相当于增加了一个空间维度,如图6右图所示,构成了一个单隐层的神经网络,决策计算就能生成两条直线,可以区别异或问题了。
To solve the problem of non-linear separability, consider using a multi-layer functional neural network. A layer of neurons between the output layer and the input layer is called the hidden layer. Both the hidden layer and the output layer neurons are functional neurons with activation functions. In the left diagram of Figure 7, the perceptron's neurons have no hidden layer, and the decision calculation only generates a straight line, which cannot distinguish the XOR problem. However, adding a hidden layer with a non-linear activation function can solve this. The non-linearity of the activation function of the hidden layer output also helps to solve the non-linear separability problem. Adding one more hidden layer is equivalent to adding a spatial dimension, as shown in the right diagram of Figure 6, forming a single hidden layer neural network, where the decision calculation can generate two straight lines, thus distinguishing the XOR problem.
图7:加一个隐层解决感知机的异或问题 Figure 7: Adding a hidden layer to solve the perceptron's XOR problem
对多隐层的神经网络,还有一个“万能近似定理”,意味着使用S型函数作为激励函数的多层神经网络,可以用来近似任意的复杂函数,并且可以达到任意近似精准度。
For multi-hidden layer neural networks, there is also a "universal approximation theorem," which means that multi-layer neural networks using sigmoid functions as activation functions can be used to approximate any complex function and can achieve any desired level of approximation accuracy.
总而言之,从20世纪80年代和90年代开始,联结主义重新出现,神经网络研究回归主流。许多人认为罗森布拉特的理论已被证明是正确的。朴素的感知器有其缺陷,但它的基本原理引发了现代人工智能革命。深度学习和神经网络正在改变我们的社会,了解一下感知器,以及神经网络这段兴衰史,有助于我们更好地认清AI,以及AI发展的未来。
In summary, starting from the 1980s and 1990s, connectionism re-emerged, and neural network research returned to the mainstream. Many people believe that Rosenblatt's theory has been proven correct. The naive perceptron has its flaws, but its basic principles sparked the modern artificial intelligence revolution. Deep learning and neural networks are transforming our society. Understanding the perceptron and the rise and fall of neural networks helps us better recognize AI and the future of AI development.
参考文献:(上下滑动可浏览)
References:(Scroll up and down to browse)
Scan to Follow