论文标题

表征不同学习模型的重量空间

Characterizing the Weight Space for Different Learning Models

论文作者

Musunuru, Saurav, Paranjape, Jay N., Dubey, Rahul Kumar, Venkoparao, Vijendran G.

论文摘要

深度学习已成为开发智能机器的主要研究领域之一。 AI的大多数著名应用(例如语音识别,图像处理和NLP)都是由深度学习驱动的。深度学习算法使用人工神经网络模仿人脑,并逐步学会准确解决给定的问题。但是深度学习系统面临重大挑战。有很多尝试使深度学习模型模仿生物神经网络。但是,在存在对抗性例子的情况下,许多深度学习模型的表现较差。对抗性例子的表现不佳会导致对抗性攻击,进而导致大多数应用程序的安全性和安全性。在本文中,我们试图以三种不同的子集的形式来表征深神经网络的解决方案空间。属于精确训练的图案的权重,属于广义图案集的权重和属于对抗模式集的权重。我们试图用两个看似不同的学习范式来表征解决方案空间。深度神经网络和密集的关联记忆模型,它们试图通过完全不同的机制实现学习。我们还表明,对抗性攻击通常不如深度神经网络对关联记忆模型的成功率不大。

Deep Learning has become one of the primary research areas in developing intelligent machines. Most of the well-known applications (such as Speech Recognition, Image Processing and NLP) of AI are driven by Deep Learning. Deep Learning algorithms mimic human brain using artificial neural networks and progressively learn to accurately solve a given problem. But there are significant challenges in Deep Learning systems. There have been many attempts to make deep learning models imitate the biological neural network. However, many deep learning models have performed poorly in the presence of adversarial examples. Poor performance in adversarial examples leads to adversarial attacks and in turn leads to safety and security in most of the applications. In this paper we make an attempt to characterize the solution space of a deep neural network in terms of three different subsets viz. weights belonging to exact trained patterns, weights belonging to generalized pattern set and weights belonging to adversarial pattern sets. We attempt to characterize the solution space with two seemingly different learning paradigms viz. the Deep Neural Networks and the Dense Associative Memory Model, which try to achieve learning via quite different mechanisms. We also show that adversarial attacks are generally less successful against Associative Memory Models than Deep Neural Networks.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源