图深度学习-第2部分

有关深层学习的FAU讲义 (FAU LECTURE NOTES ON DEEP LEARNING)

These are the lecture notes for FAU’s YouTube Lecture “Deep Learning”. This is a full transcript of the lecture video & matching slides. We hope, you enjoy this as much as the videos. Of course, this transcript was created with deep learning techniques largely automatically and only minor manual modifications were performed. Try it yourself! If you spot mistakes, please let us know!

这些是FAU YouTube讲座“ 深度学习 ”的 讲义 这是演讲视频和匹配幻灯片的完整记录。 我们希望您喜欢这些视频。 当然,此成绩单是使用深度学习技术自动创建的,并且仅进行了较小的手动修改。 自己尝试! 如果发现错误,请告诉我们!

导航 (Navigation)

Previous Lecture / Watch this Video / Top Level / Next Lecture

上一个讲座 / 观看此视频 / 顶级 / 下一个讲座

Image for post
Graph deep learning and physical simulation go well together. Image created using gifify. Source: YouTube.
图深度学习和物理模拟结合得很好。 使用gifify创建的图像 。 资料来源: YouTube 。

Welcome back to deep learning. So today, we want to continue talking about graph convolutions. We will look into the second part where we now see whether we have to stay in this spectral domain or whether we can also go back to the spatial domain. So let’s look at what I have for you.

欢迎回到深度学习。 所以今天,我们要继续讨论图卷积。 我们将研究第二部分,现在我们看看是必须保留在此频谱域中还是必须回到空间域。 因此,让我们看看我为您准备的。

Image for post
CC BY 4.0 from the 深度学习讲座中 Deep Learning Lecture.CC BY 4.0下的图像。

Remember we had this polynomial to define a convolution in the spectral domain. We’ve seen that by computing the eigenvectors of the Laplacian matrix, we were able to find an appropriate Fourier transform that would then give us a spectral representation of the graph configuration. Then, we could do our convolution in the spectral domain and transform it back. Now, this was kind of very expensive because we have to compute U. For U, we have to do the eigenvalue decomposition for this entire symmetric matrix. Also, we’ve seen that we can’t use tricks of the fast Fourier transform because this doesn’t necessarily hold for our U.

记住我们有这个多项式来定义谱域中的卷积。 我们已经看到,通过计算拉普拉斯矩阵的特征向量,我们能够找到合适的傅立叶变换,然后该傅立叶变换将为我们提供图配置的频谱表示。 然后,我们可以在频谱域中进行卷积并将其转换回去。 现在,这非常昂贵,因为我们必须计算U。 对于U,我们必须对整个对称矩阵进行特征值分解。 另外,我们已经看到我们不能使用快速傅立叶变换的技巧,因为这不一定适用于我们的U。

So, how can we choose now our k and θ in order to get rid of U? Well, so if we choose k equals to 1, θ subscript 0 to 2θ, and θ subscript 1 to -θ, we get the following polynomial. So, we still have the configuration that we have x transformed in the Fourier space times our polynomial expressed as matrix times the inverse Fourier transform here. Now, let’s look into the configuration of G hat. G hat can actually be expressed as 2 times θ times Λ to the power of 0. Remember Λ is a diagonal matrix. So, we take every element to the power of 0. This is actually a unity matrix and we subtract θ times Λ to the power of 1. Well, this is actually just Λ. Then, we can express our complete matrix G hat in this way. Of course, we can then pull in our U from the left-hand side and the right-hand side which is giving us the following expression. Now, we use the property that θ is actually a scalar. So, we can pull it to the front. The Λ to the power of 0 cancels out because this is essentially an identity matrix. The Λ on the right-hand side term still remains, but we can also pull out the θ. Well the UU transpose just cancels out. So, this is again the identity matrix and we can use our definition of the symmetric version of our graph Laplacian. You can see that we’ve just found it, here in our equation. So, we can also replace it with this one. You see now U is suddenly gone. So, we can pull out θ again and all that remains is that we have two times the identity matrix minus the symmetric version of the graph Laplacian. If we now plug in the definition of the symmetric version associated with the original adjacency matrix and the degree matrix, we can see that we still can plug this definition in. Then, one of the identity matrices cancels out and we finally get identity plus D to the power of -0.5 times Atimes D to the power of -0.5. So, remember D is a diagonal matrix. We can easily invert the elements on the diagonal and we can also take element-wise the square root. So, this is perfectly fine. This way we don’t have U at all coming up here. We can express our entire graph convolution in this very nice way using the graph Laplacian matrix.

那么,我们现在如何选择k和θ来摆脱U呢? 好,因此,如果我们选择k等于1,θ下标0到2θ,θ下标1到-θ,我们得到以下多项式。 因此,我们仍然具有以下配置:在傅立叶空间中将x变换为乘以矩阵表示的多项式乘以此处的傅立叶逆变换。 现在,让我们研究一下G hat的配置。 如2倍θ倍λ〜0的功率记住Λ是对角矩阵G帽子实际上可以表达。 因此,我们将每个元素取0的幂。这实际上是一个单位矩阵,我们将θ乘以Λ乘以1的幂。好吧,这实际上只是Λ 。 然后,我们可以用这种方式表示完整的矩阵G hat。 当然,然后我们可以从左侧和右侧拉入U ,这使我们得到以下表达式。 现在,我们使用θ实际上是一个标量的属性。 因此,我们可以将其拉到最前面。 幂为0的Λ被抵消,因为这本质上是一个单位矩阵。 右侧项中的Λ仍然保留,但我们也可以拉出θ。 好吧, UU转置只是抵消了。 因此,这再次是单位矩阵,我们可以使用我们的图拉普拉斯算子的对称版本的定义。 您可以在方程式中看到我们刚刚找到它。 因此,我们也可以用此替代它。 您现在看到U突然消失了。 因此,我们可以再次拉出θ,剩下的就是我们有两倍的单位矩阵减去图拉普拉斯图的对称形式。 如果现在插入与原始邻接矩阵和度矩阵关联的对称版本的定义,我们可以看到仍然可以插入此定义。然后,一个标识矩阵被抵消,最终得到标识加D达到-0.5的乘方A乘以D的-0.5的乘方。 因此,请记住D是对角矩阵。 我们可以轻松地将对角线上的元素反转,也可以对元素取平方根。 所以,这很好。 这样,我们这里根本没有U。 我们可以使用图拉普拉斯矩阵以这种非常好的方式表示整个图卷积。

Image for post
CC BY 4.0 from the 深度学习讲座中 Deep Learning Lecture.CC BY 4.0下的图像。

Now let’s analyze this term a little more. So, we can see this identity on the left-hand side, we see we can convolve in the spectral domain, and we can construct G hat as a polynomial of Laplacian filters. Then, we can see with a particular choice k equals 1, θ subscript 0 equals to 2θ and θ subscript 1 equals to -θ. Then, this term suddenly only depends on the scalar value θ. With all these tricks, we got rid of the Fourier transform U transpose. So, we suddenly can express graph convolutions in this simplified way.

现在让我们再分析一下这个术语。 因此,我们可以在左侧看到该标识,可以在谱域中进行卷积,并且可以将G hat构造为Laplacian滤波器的多项式。 然后,我们可以看到在特定选择下k等于1,θ下标0等于2θ,θ下标1等于-θ。 然后,该项突然仅取决于标量值θ。 通过所有这些技巧,我们摆脱了傅立叶变换U转置。 因此,我们突然可以用这种简化的方式表示图卷积。

Image for post
CC BY 4.0 from the 深度学习讲座中 Deep Learning Lecture.CC BY 4.0下的图像。

Well, this is the basic graph convolutional operation and you can find this actually shown in reference [1]. You can essentially do this to scalar values, you use your degree matrix and plug it in here. You use your adjacency matrix and you plug it in here. Then, you can optimize with respect to θ in order to find the weight for your convolutions.

好吧,这是基本的图卷积运算,您可以在参考文献[1]中找到它。 您基本上可以对标量值执行此操作,您可以使用度矩阵并将其插入此处。 您使用邻接矩阵,并将其插入此处。 然后,您可以针对θ进行优化,以找到卷积的权重。

Image for post
CC BY 4.0 from the 深度学习讲座中 Deep Learning Lecture.CC BY 4.0下的图像。

Well, now the question is “Is it really necessary to motivate the graph convolution from the spectral domain?” and the answer is “No.”. So, we can also motivate it spatially.

好吧,现在的问题是“真的有必要从谱域中激发图卷积吗?” 答案是“否”。 因此,我们还可以在空间上激发它。

Image for post
CC BY 4.0 from the 深度学习讲座中 Deep Learning Lecture.CC BY 4.0下的图像。

Well, let’s look at the following concept. For a mathematician, a graph is a manifold, but a discrete one. We can discretize the manifold and do a spectral convolution using the Laplacian matrix. So, this led us to spectral graph convolutions. But as a computer scientist, you can interpret a graph as a set of nodes and vertices connected through edges. We now need to define how to aggregate the information of one vertex through its neighbors. If we do so, we get the spatial graph convolution.

好吧,让我们看一下以下概念。 对于数学家来说,图是流形,但是离散的。 我们可以离散流形,并使用拉普拉斯矩阵进行频谱卷积。 因此,这导致我们进行频谱图卷积。 但是作为计算机科学家,您可以将图解释为通过边连接的一组节点和顶点。 现在,我们需要定义如何通过一个顶点的邻居聚合一个顶点的信息。 如果这样做,我们将获得空间图卷积。

Image for post
CC BY 4.0 from the 深度学习讲座中 Deep Learning Lecture.CC BY 4.0下的图像。

Well, how is this done? One approach shown in [2] is GraphSAGE. Here, we essentially define a vertex of interest and we define how neighbors contribute to the vertex of interest. So technically, we implement this using a feature vector at the node v and the k-th layer. This can be described as h k subscript v. So, for the zeroth layer, this contains the input. This is just the original configuration of your graph. Then, we need to be able to aggregate in order to compute the next layer. This is done by a spatial aggregation function over the previous layer. Therefore, you use all of the neighbors and typically you define this neighborhood such that every node that is connected to the node under consideration is included in this neighborhood.

好吧,这是怎么做的? [2]中显示的一种方法是GraphSAGE。 在这里,我们本质上定义了关注的顶点,并且定义了邻居如何对关注的顶点做出贡献。 因此,从技术上讲,我们在节点v和第k层使用特征向量来实现这一点。 这可以描述为h k下标v。因此,对于第0层,它包含输入。 这只是图形的原始配置。 然后,我们需要能够聚合以便计算下一层。 这是通过上一层的空间聚合功能完成的。 因此,您将使用所有邻居,并且通常会定义该邻域,以便连接到所考虑节点的每个节点都包含在此邻域中。

Image for post
CC BY 4.0 from the 深度学习讲座中 Deep Learning Lecture.CC BY 4.0下的图像。

So this line brings us to the GraphSAGE algorithm. Here, you start with a graph and input features. Then, you do the following algorithm: You initialize at h 0 with simply the input of the graph configuration. Then, you iterate over the layers. You iterate over the nodes. For every node, you run the aggregation function that somehow computes a summary over all of your neighbors. Then, the result is a vector of a certain dimension and you then take the aggregated vector and the current configuration of the vector, you concatenate them and multiply them with a weight matrix. This is then run through a non-linearity. Lastly, you scale by the magnitude of your activations. This is then iterated over all of the layers and finally, you get the output z that is the result of your graph convolution.

因此,这条线将我们带到了GraphSAGE算法。 在这里,您将从图形和输入要素开始。 然后,执行以下算法:仅使用图配置的输入在h 0进行初始化。 然后,您遍历各层。 您遍历节点。 对于每个节点,您都可以运行聚合函数,以某种方式计算所有邻居的汇总。 然后,结果是一个特定维度的向量,然后取聚合向量和向量的当前配置,将它们连接起来,然后将它们与权重矩阵相乘。 然后,这通过非线性进行。 最后,您可以根据激活的大小进行扩展。 然后在所有层上进行迭代,最后得到图卷积结果的输出z

Image for post
CC BY 4.0 from the 深度学习讲座中 Deep Learning Lecture.CC BY 4.0下的图像。

The concept of aggregators is key to develop this algorithm because in every node you may have a different number of neighbors. A very simple aggregator would then be simply computing the mean. Of course, you can also take the GCN aggregator and that brings us back to the spectral representation. This way, the connection between spatial and spectral domains can be established. Furthermore, you can take a pooling aggregator which then uses, for example, maximum pooling or you use recurrent networks like LSTM aggregators.

聚合器的概念是开发此算法的关键,因为在每个节点中,您可能具有不同数量的邻居。 然后,一个非常简单的聚合器将简单地计算均值。 当然,您也可以使用GCN聚合器,这使我们回到了光谱表示形式。 这样,可以建立空间域和光谱域之间的连接。 此外,您可以使用一个池聚合器,然后使用例如最大池化,或者使用像LSTM聚合器这样的循环网络。

Image for post
CC BY 4.0 from the 深度学习讲座中 Deep Learning Lecture.CC BY 4.0下的图像。

You already see that there is a broad variety of aggregators. This then also is the reason why there are so many different graph deep learning approaches. You can subdivide them into certain kinds because there are spectral ones, there are spatial ones, and there are the recurrent ones. So, this is essentially the key how you can tackle the graph convolutional neural networks. So, what do we actually want to do? Well, you can take one of these algorithms and apply it to some mesh. Of course, this can also be done on very complex meshes and I will put a couple of references below that you can see what kind of applications can be done. For example, you can use these methods in order to process the information on coronary arteries.

您已经看到了各种各样的聚合器。 这也是为什么有这么多不同的图深度学习方法的原因。 您可以将它们细分为某些种类,因为有频谱种类,空间种类以及循环种类。 因此,这实际上是解决图卷积神经网络的关键。 那么,我们实际上想做什么? 好了,您可以采用这些算法之一,并将其​​应用于某些网格。 当然,这也可以在非常复杂的网格上完成,我将在下面放置一些参考,以了解可以完成哪种应用程序。 例如,您可以使用这些方法来处理冠状动脉信息。

Image for post
CC BY 4.0 from the 深度学习讲座中 Deep Learning Lecture.CC BY 4.0下的图像。

Well next time in deep learning, there’s only a couple of topics left. One thing that I want to show to you is how you can embed prior knowledge into deep networks. This is also a quite nice idea because it allows us to fuse much of the things that we know from theory and signal processing with our deep learning approaches. Of course, I also have a couple of references and if you have some time please read through them. They elaborate much more closely on the ideas that we presented here. There are also image references that I’ll put into the description of this video. So, thank you very much for listening and see you in the next lecture. Bye-bye!

下次在深度学习中,只剩下了几个主题。 我想向您展示的一件事是如何将先验知识嵌入到深层网络中。 这也是一个很好的主意,因为它使我们能够将我们从理论和信号处理中了解到的许多知识与我们的深度学习方法融合在一起。 当然,我也有一些参考资料,如果您有时间请仔细阅读。 他们更加详细地阐述了我们在此处提出的想法。 我还将在本视频的说明中加入图像参考。 因此,非常感谢您的收听,并在下一堂课中与您相见。 再见!

Image for post
Many more important concepts had to be omitted here. Therefore enjoy further reading on Graph Deep Learning below. Image created using gifify. Source: YouTube.
这里必须省略许多更重要的概念。 因此,请享受下面的图深度学习的进一步阅读。 使用gifify创建的图像 。 资料来源: YouTube 。

If you liked this post, you can find more essays here, more educational material on Machine Learning here, or have a look at our Deep LearningLecture. I would also appreciate a follow on YouTube, Twitter, Facebook, or LinkedIn in case you want to be informed about more essays, videos, and research in the future. This article is released under the Creative Commons 4.0 Attribution License and can be reprinted and modified if referenced. If you are interested in generating transcripts from video lectures try AutoBlog.

如果你喜欢这篇文章,你可以找到这里更多的文章 ,更多的教育材料,机器学习在这里 ,或看看我们的深入 学习 讲座 。 如果您希望将来了解更多文章,视频和研究信息,也欢迎关注YouTube , Twitter , Facebook或LinkedIn 。 本文是根据知识共享4.0署名许可发布的 ,如果引用,可以重新打印和修改。 如果您对从视频讲座中生成成绩单感兴趣,请尝试使用AutoBlog 。

谢谢 (Thanks)

Many thanks to the great introduction by Michael Bronstein on MISS 2018! and special thanks to Florian Thamm for preparing this set of slides.

非常感谢Michael Bronstein在MISS 2018上的精彩介绍! 特别感谢Florian Thamm准备了这组幻灯片。

翻译自: https://towardsdatascience.com/graph-deep-learning-part-2-c6110d49e63c

本文来自互联网用户投稿,该文观点仅代表作者本人,不代表本站立场。本站仅提供信息存储空间服务,不拥有所有权,不承担相关法律责任。如若转载,请注明出处:http://www.mzph.cn/news/390809.shtml

如若内容造成侵权/违法违规/事实不符,请联系多彩编程网进行投诉反馈email:809451989@qq.com,一经查实,立即删除!

相关文章

Linux下 安装Redis并配置服务

一、简介 1、 Redis为单进程单线程模式,采用队列模式将并发访问变成串行访问。 2、 Redis不仅仅支持简单的k/v类型的数据,同时还提供list,set,zset,hash等数据结构的存储。 3、 Redis支持数据的备份,即mas…

leetcode 477. 汉明距离总和(位运算)

theme: healer-readable 题目 两个整数的 汉明距离 指的是这两个数字的二进制数对应位不同的数量。 计算一个数组中,任意两个数之间汉明距离的总和。 示例: 输入: 4, 14, 2 输出: 6 解释: 在二进制表示中,4表示为0100,14表示为1110&…

量子信息与量子计算_量子计算为23美分。

量子信息与量子计算On Aug 13, 2020, AWS announced the General Availability of Amazon Braket. Braket is their fully managed quantum computing service. It is available on an on-demand basis, much like SageMaker. That means the everyday developer and data scie…

全面理解Java内存模型

Java内存模型即Java Memory Model,简称JMM。JMM定义了Java 虚拟机(JVM)在计算机内存(RAM)中的工作方式。JVM是整个计算机虚拟模型,所以JMM是隶属于JVM的。 如果我们要想深入了解Java并发编程,就要先理解好Java内存模型。Java内存模型定义了多…

leetcode 1074. 元素和为目标值的子矩阵数量(map+前缀和)

给出矩阵 matrix 和目标值 target&#xff0c;返回元素总和等于目标值的非空子矩阵的数量。 子矩阵 x1, y1, x2, y2 是满足 x1 < x < x2 且 y1 < y < y2 的所有单元 matrix[x][y] 的集合。 如果 (x1, y1, x2, y2) 和 (x1’, y1’, x2’, y2’) 两个子矩阵中部分坐…

失物招领php_新奥尔良圣徒队是否增加了失物招领?

失物招领phpOver the past couple of years, the New Orleans Saints’ offense has been criticized for its lack of wide receiver options. Luckily for Saints’ fans like me, this area has been addressed by the signing of Emmanuel Sanders back in March — or has…

leetcode 5756. 两个数组最小的异或值之和(状态压缩dp)

题目 给你两个整数数组 nums1 和 nums2 &#xff0c;它们长度都为 n 。 两个数组的 异或值之和 为 (nums1[0] XOR nums2[0]) (nums1[1] XOR nums2[1]) … (nums1[n - 1] XOR nums2[n - 1]) &#xff08;下标从 0 开始&#xff09;。 比方说&#xff0c;[1,2,3] 和 [3,2,1…

客户细分模型_Avarto金融解决方案的客户细分和监督学习模型

客户细分模型Lets assume that you are a CEO of a company which have some X amount of customers in a city with 1000 *X population. Analyzing the trends/features of your customer and segmenting the population of the city to land new potential customers would …

leetcode 231. 2 的幂

给你一个整数 n&#xff0c;请你判断该整数是否是 2 的幂次方。如果是&#xff0c;返回 true &#xff1b;否则&#xff0c;返回 false 。 如果存在一个整数 x 使得 n 2x &#xff0c;则认为 n 是 2 的幂次方。 示例 1&#xff1a; 输入&#xff1a;n 1 输出&#xff1a;tr…

leetcode 342. 4的幂

给定一个整数&#xff0c;写一个函数来判断它是否是 4 的幂次方。如果是&#xff0c;返回 true &#xff1b;否则&#xff0c;返回 false 。 整数 n 是 4 的幂次方需满足&#xff1a;存在整数 x 使得 n 4x 示例 1&#xff1a; 输入&#xff1a;n 16 输出&#xff1a;true …

梯度反传_反事实政策梯度解释

梯度反传Among many of its challenges, multi-agent reinforcement learning has one obstacle that is overlooked: “credit assignment.” To explain this concept, let’s first take a look at an example…在许多挑战中&#xff0c;多主体强化学习有一个被忽略的障碍&a…

大数据与Hadoop

大数据的定义 大数据是指无法在一定时间内用常规软件工具对其内容进行抓取、管理和处理的数据集合。 大数据的概念–4VXV 1,数据量大&#xff08;Volume&#xff09;2,类型繁多&#xff08;Variety &#xff09;3,速度快时效高&#xff08;Velocity&#xff09;4,价值密度低…

facebook.com_如何降低电子商务的Facebook CPM

facebook.comWith the 2020 election looming, Facebook advertisers and e-commerce stores are going to continually see their ad costs go up as the date gets closer (if they haven’t already).随着2020年选举的临近&#xff0c;随着日期越来越近&#xff0c;Facebook…

Hadoop安装及配置

Hadoop的三种运行模式 单机模式&#xff08;Standalone,独立或本地模式&#xff09;:安装简单,运行时只启动单个进程,仅调试用途&#xff1b;伪分布模式&#xff08;Pseudo-Distributed&#xff09;:在单节点上同时启动namenode、datanode、secondarynamenode、resourcemanage…

漏洞发布平台-安百科技

一个不错的漏洞发布平台&#xff1a;https://vul.anbai.com/ 转载于:https://blog.51cto.com/antivirusjo/2093758

西格尔零点猜想_我从埃里克·西格尔学到的东西

西格尔零点猜想I finished reading Eric Siegel’s Predictive Analytics. And I have to say it was an awesome read. How do I define an awesome or great book? A book that changes your attitude permanently. You must not be the same person that you were before y…

HDFS 技术

HDFS定义 Hadoop Distributed File System&#xff0c;是一个使用 Java 实现的、分布式的、可横向扩展的文件系 统&#xff0c;是 HADOOP 的核心组件 HDFS特点 处理超大文件流式地访问数据运行于廉价的商用机器集群上&#xff1b; HDFS 不适合以下场合&#xff1a;低延迟数据…

深度学习算法和机器学习算法_啊哈! 4种流行的机器学习算法的片刻

深度学习算法和机器学习算法Most people are either in two camps:大多数人都在两个营地中&#xff1a; I don’t understand these machine learning algorithms. 我不了解这些机器学习算法。 I understand how the algorithms work, but not why they work. 我理解的算法是如…

Python第一次周考(0402)

2019独角兽企业重金招聘Python工程师标准>>> 一、单选 1、Python3中下列语句错误的有哪些&#xff1f; A s input() B s raw_input() C print(hello world.) D print(hello world.) 2、下面哪个是 Pycharm 在 Windows 下 默认 用于“批量注释”的快捷键 A Ctrl d…

ASP.NET 页面之间传值的几种方式

对于任何一个初学者来说&#xff0c;页面之间传值可谓是必经之路&#xff0c;却又是他们的难点。其实&#xff0c;对大部分高手来说&#xff0c;未必不是难点。 回想2016年面试的将近300人中&#xff0c;有实习生&#xff0c;有应届毕业生&#xff0c;有1-3年经验的&#xff0c…