diff --git "a/txt/2105.00696.txt" "b/txt/2105.00696.txt" new file mode 100644--- /dev/null +++ "b/txt/2105.00696.txt" @@ -0,0 +1,2070 @@ +IEEE TRANSACTIONS ON ARTIFICIAL INTELLIGENCE, VOL. 00, NO. 0, 2021 1 +Graph Learning: A Survey +Feng Xia, Senior Member, IEEE, Ke Sun, Shuo Yu, Member, IEEE, Abdul Aziz, Liangtian Wan, Member, IEEE, +Shirui Pan, and Huan Liu, Fellow, IEEE +Abstract —Graphs are widely used as a popular representation +of the network structure of connected data. Graph data can +be found in a broad spectrum of application domains such +as social systems, ecosystems, biological networks, knowledge +graphs, and information systems. With the continuous penetra- +tion of artificial intelligence technologies, graph learning (i.e., +machine learning on graphs) is gaining attention from both +researchers and practitioners. Graph learning proves effective for +many tasks, such as classification, link prediction, and matching. +Generally, graph learning methods extract relevant features of +graphs by taking advantage of machine learning algorithms. +In this survey, we present a comprehensive overview on the +state-of-the-art of graph learning. Special attention is paid to +four categories of existing graph learning methods, including +graph signal processing, matrix factorization, random walk, +and deep learning. Major models and algorithms under these +categories are reviewed respectively. We examine graph learning +applications in areas such as text, images, science, knowledge +graphs, and combinatorial optimization. In addition, we discuss +several promising research directions in this field. +Index Terms —Graph learning, graph data, machine learning, +deep learning, graph neural networks, network representation +learning, network embedding. +IMPACT STATEMENT +Real-world intelligent systems generally rely on machine +learning algorithms handling data of various types. Despite +their ubiquity, graph data have imposed unprecedented chal- +lenges to machine learning due to their inherent complexity. +Unlike text, audio and images, graph data are embedded +in an irregular domain, making some essential operations +of existing machine learning algorithms inapplicable. Many +graph learning models and algorithms have been developed +to tackle these challenges. This paper presents a systematic +review of the state-of-the-art graph learning approaches as +well as their potential applications. The paper serves mul- +tiple purposes. First, it acts as a quick reference to graph +learning for researchers and practitioners in different areas +such as social computing, information retrieval, computer +vision, bioinformatics, economics, and e-commence. Second, +it presents insights into open areas of research in the field. +Third, it aims to stimulate new research ideas and more +interests in graph learning. +I. I NTRODUCTION +F. Xia is with School of Engineering, IT and Physical Sciences, Federation +University Australia, Ballarat, VIC 3353, Australia +K. Sun, S. Yu, A. Aziz, and L. Wan are with School of Software, Dalian +University of Technology, Dalian 116620, China. +S. Pan is with Faculty of Information Technology, Monash University, +Melbourne, VIC 3800, Australia. +H. Liu is with School of Computing, Informatics, and Decision Systems +Engineering, Arizona State University, Tempe, AZ 85281, USA. +Corresponding author: Feng Xia; e-mail: f.xia@ieee.orgGRAPHS, also referred to as networks, can be extracted +from various real-world relations among abundant enti- +ties. Some common graphs have been widely used to formulate +different relationships, such as social networks, biological +networks, patent networks, traffic networks, citation networks, +and communication networks [1]–[3]. A graph is often defined +by two sets, i.e., vertex set and edge set. Vertices represent en- +tities in graph, whereas edges represent relationships between +those entities. Graph learning has attracted considerable atten- +tion because of its wide applications in the real world, such as +data mining and knowledge discovery. Graph learning meth- +ods have gained increasing popularity for capturing complex +relationships, as graphs exploit essential and relevant relations +among vertices [4], [5]. For example, in microblog networks, +the spread trajectory of rumors can be tracked by detecting +information cascades. In biological networks, new treatments +for difficult diseases can be discovered by inferring protein +interactions. In traffic networks, human mobility patterns can +be predicted by analyzing the co-occurrence phenomenon with +different timestamps [6]. Efficient analysis of these networks +massively depends on the way how networks are represented. +A. What is Graph Learning? +Generally speaking, graph learning refers to machine learn- +ing on graphs. Graph learning methods map the features +of a graph to feature vectors with the same dimensions in +the embedding space. A graph learning model or algorithm +directly converts the graph data into the output of the graph +learning architecture without projecting the graph into a low +dimensional space. Most graph learning methods are based +on or generalized from deep learning techniques, because +deep learning techniques can encode and represent graph data +into vectors. The output vectors of graph learning are in +continuous space. The target of graph learning is to extract +the desired features of a graph. Thus the representation of +a graph can be easily used by downstream tasks such as +node classification and link prediction without an explicit +embedding process. Consequently, graph learning is a more +powerful and meaningful technique for graph analysis. +In this survey paper, we try to examine machine learning +methods on graphs in a comprehensive manner. As shown +in Fig. 1, we focus on existing methods that fall into the +following four categories: graph signal processing (GSP) based +methods, matrix factorization based methods, random walk +based methods, and deep learning based methods. Roughly +speaking, GSP deals with sampling and recovery of graph, and +learning topology structure from data. Matrix factorization can +be divided into graph Laplacian matrix factorization and vertex +proximity matrix factorization. Random walk based methodsarXiv:2105.00696v1 [cs.LG] 3 May 2021IEEE TRANSACTIONS ON ARTIFICIAL INTELLIGENCE, VOL. 00, NO. 0, 2021 2 +Fig. 1: The categorization of graph learning. +include structure-based random walk, structure and node in- +formation based random walk, random walk in heterogeneous +networks, and random walk in time-varying networks. Deep +learning based methods include graph convolutional networks, +graph attention networks, graph auto-encoder, graph generative +networks, and graph spatial-temporal networks. Basically, the +model architectures of these methods/techniques differ from +each other. This paper presents an extensive review of the +state-of-the-art graph learning techniques. +Traditionally, researchers adopt an adjacency matrix to +represent a graph, which can only capture the relationship +between two adjacent vertices. However, many complex and +irregular structures cannot be captured by this simple rep- +resentation. When we analyze large-scale networks, tradi- +tional methods are computationally expensive and hard to be +implemented in real-world applications. Therefore, effective +representation of these networks is a paramount problem to +solve [4]. Network Representation Learning (NRL) proposed +in recent years can learn latent features of network vertices +with low dimensional representation [7]–[9]. When the new +representation has been learned, previous machine learning +methods can be employed for analyzing the graph data as well +as discovering relationships hidden in the data. +When complex networks are embedded into a latent, low +dimensional space, the structural information and vertex at- +tributes can be preserved [4]. Thus the vertices of networks can +be represented by low dimensional vectors. These vectors can +be regarded as the features of input in previous machine learn- +ing methods. Graph learning methods pave the way for graph +analysis in the new representation space, and many graph +analytical tasks, such as link prediction, recommendation and +classification, can be solved efficiently [10], [11]. Graphical +network representation sheds light on various aspects of social +life, such as communication patterns, community structure, +and information diffusion [12], [13]. According to the at- +tributes of vertices, edges and subgraph, graph learning tasks +can be divided into three categories, which are vertices based, +edges based, and subgraph based, respectively. The relation- +ships among vertices in a graph can be exploited for, e.g.,classification, risk identification, clustering, and community +detection [14]. By judging the presence of edges between +two vertices in graphs, we can perform recommendation and +knowledge reasoning, for instance. Based on the classification +of subgraphs [15], the graph can be used for, e.g., polymer +classification, 3D visual classification, etc. For GSP, it is sig- +nificant to design suitable graph sampling methods to preserve +the features of the original graph, which aims at recovering the +original graph efficiently [16]. Graph recovery methods can +be used for constructing the original graph in the presence +of incomplete data [17]. Afterwards, graph learning can be +exploited to learn the topology structure from graph data. In +summary, graph learning can be used to tackle the following +challenges, which are difficult to solve by using traditional +graph analysis methods [18]. +1)Irregular domains: Data collected by traditional sen- +sors have a clear grid structure. However, graphs lie in an +irregular domain (i.e., non-Euclidean space). In contrast +to regular domain (i.e., Euclidean space), data in non- +Euclidean space are not ordered regularly. Distance is +hence difficult to be defined. As a result, basic methods +based on traditional machine learning and signal pro- +cessing cannot be directly generalized to graphs. +2)Heterogeneous networks: In many cases, networks +involved in the traditional graph analysis algorithms +are homogeneous. The appropriate modeling methods +only consider the direct connection of the network and +strip other irrelevant information, which significantly +simplifies the processing. However, it is prone to cause +information loss. In the real world, the edges among +vertices and the types of vertices are usually diverse, +such as in the academic network shown in Fig. 2. Thus it +isn’t easy to discover potential value from heterogeneous +information networks with abundant vertices and edges. +3)Distributed algorithms: In big social networks, there +are often millions of vertices and edges [19]. Centralized +algorithms cannot handle this since the computational +complexity of these algorithms would significantly in- +crease with the growth of vertex number. The design ofIEEE TRANSACTIONS ON ARTIFICIAL INTELLIGENCE, VOL. 00, NO. 0, 2021 3 +distributed algorithms for dealing with big networks is a +critical problem yet to be solved [20]. One major benefit +of distributed algorithms is that the algorithms can be +executed in multiple CPUs or GPUs simultaneously, and +hence the running time can be reduced significantly. +B. Related Surveys +There are several surveys that are partially related to the +scope of this paper. Unlike these surveys, we aim to provide +a comprehensive overview of graph learning methods, with a +focus on four specific categories. In particular, graph signal +processing is introduced as one approach for graph learning, +which is not covered by other surveys. +Goyal and Ferrara [21] summarized graph embedding meth- +ods, such as matrix factorization, random walk and their +applications in graph analysis. Cai et al. [22] reviewed graph +embedding methods based on problem settings and embedding +techniques. Zhang et al. [4] summarized NRL methods based +on two categories, i.e., unsupervised NRL and semi-supervised +NRL, and discussed their applications. Nickel et al. [23] +introduced knowledge extraction methods from two aspects: +latent feature models and graph based models. Akoglu et +al. [24] reviewed state-of-the-art techniques for event detec- +tion in data represented as graphs, and their applications in +the real world. Zhang et al. [18] summarized deep learning +based methods for graphs, such as graph neural networks +(GNNs), graph convolutional networks (GCNs) and graph +auto-encoders (GAEs). Wu et al. [25] reviewed state-of-the- +art GNN methods and discussed their applications in dif- +ferent fields. Ortega et al. [26] introduced GSP techniques +for representation, sampling and learning, and discussed their +applications. Huang et al. [27] examined the applications of +GSP in functional brain imaging and addressed the problem of +how to perform brain network analysis from signal processing +perspective. +In summary, none of the existing surveys provides a com- +prehensive overview of graph learning. They only cover some +parts of graph learning, such as network embedding and deep +learning based network representation. The NRL and/or GNN +based surveys do not cover the GSP techniques. In contrast, we +review GSP techniques in the context of graph learning, as it +is an important approach for GNNs. Specifically, this survey +paper integrates state-of-the-art machine learning techniques +for graph data, gives a general description of graph learning, +and discusses its applications in various domains. +C. Contributions and Organization +The contributions of this paper can be summarized as +follows. +A comprehensive overview of state-of-the-art graph +learning methods: we present an integral introduction +to graph learning methods, including, e.g., technical +sketches, application scenarios, and potential research +directions. +Taxonomy of graph learning: we give a technical clas- +sification of mainstream graph learning methods from the +perspective of theoretical models. Technical descriptions +Fig. 2: Heterogeneous academic network [28]. +are provided wherever appropriate to improve understand- +ing of the taxonomy. +Insights into future directions in graph learning: +Besides qualitative analysis of existing methods, we shed +light on potential research directions in the field of graph +learning through summarizing several open issues and +relevant challenges. +The rest of this paper is organized as follows. An overview +of graph learning approaches containing graph signal pro- +cessing based methods, matrix factorization based methods, +random walk based methods, and deep learning based methods +is provided in Section II. The applications of graph learning +are examined in Section III. Some future directions as well +as challenges are discussed in Section IV. We conclude the +survey in Section V. +II. G RAPH LEARNING MODELS AND ALGORITHMS +The feature vectors that represent various categorical at- +tributes are viewed as the input in previous machine learning +methods. However, the mapping from the input feature vectors +to the output prediction results need to be handled by graph +learning [21]. Deep learning has been regarded as one of the +most successful techniques in artificial intelligence [29], [30]. +Extracting complex patterns by exploiting deep learning from +a massive amount of irregular data has been found very useful +in various fields, such as pattern recognition and image pro- +cessing. Consequently, how to utilize deep learning techniques +to extract patterns from complex graphs has attracted lots of +attention. Deep learning on graphs, such as GNNs, GCNs, +and GAEs, has been recognized as a powerful technique for +graph analysis [18]. Besides, GSP has also been proposed +to deal with graph analysis [26]. One of the most typical +scenarios is that a set of values reside on a set of vertices, +and these vertices are connected by edges [31]. Graph signals +can be adopted to model various phenomena in real world. For +example, in social networks, users in Facebook can be viewed +as vertices, and their friendships can be modeled as edges. The +number of followers of each vertex is marked in this social +network. Based on this assumption, many techniques (e.g., +convolution, filter, wavelet, etc.) in classical signal processing +can be employed for GSP with suitable modifications [26]. +In this section, we review graph learning models and algo- +rithms under four categories as mentioned before, namely GSPIEEE TRANSACTIONS ON ARTIFICIAL INTELLIGENCE, VOL. 00, NO. 0, 2021 4 +based methods, matrix factorization based methods, random +walk based methods, and deep learning based methods. In +Table I, we list the abbreviations used in this paper. +TABLE I: Definitions of abbreviations +Abbreviation Definition +PCA Principal component analysis +NRL Network representation learning +LSTM Long short-term memory (networks) +GSP Graph signal processing +GNN Graph neural network +GMRF Gauss markov random field +GCN Graph convolutional network +GAT Graph attention network +GAN Generative adversarial network +GAE Graph auto-encoder +ASP Algebraic signal processing +RNN Recurrent neural network +CNN Convolutional neural network +A. Graph Signal Processing +Signal processing is a traditional subject that processes +signals defined in regular data domain. In recent years, re- +searchers extend concepts of traditional signal processing into +graphs. Classical signal processing techniques and tools such +as Fourier transform and filtering can be used to analyze +graphs. In general, graphs are a kind of irregular data, which +are hard to handle directly. As a complement to learning +methods based on structures and models, GSP provides a new +perspective of spectral analysis of graphs. Derived from signal +processing, GSP can give an explanation of graph property +consisting of connectivity, similarity, etc. Fig. 3 gives a simple +example of graph signals at a certain time point, which is +defined as observed values. In a graph, the above mentioned +observed values can be regarded as graph signals. Each node is +then mapped to the real number field in GSP. The main task +of GSP is to expand signal processing approaches to mine +implicit information in graphs. +Fig. 3: The measurements of PM2.5 from different sensors on +July 5, 2014 (data source: https://www.epa.gov/).1) Representation on Graphs: A meaningful representation +of graphs has contributed a lot to the rapid growth of graph +learning. There are two main models of GSP, i.e., adjacency +matrix based GSP [31] and Laplacian based GSP [32]. An +adjacency matrix based GSP comes from algebraic signal +processing (ASP) [33], which interprets linear signal process- +ing from algebraic theory. Linear signal processing contains +signals, filters, signal transformation, etc. It can be applied +in both continuous and discrete time domains. The basic +assumption of linear algebra is extended to the algebra space in +ASP. By selecting signal model appropriately, ASP can obtain +different instances in linear signal processing. In adjacency +matrix based GSP, the signal model is generated from a shift. +Similar to traditional signal processing, a shift in GSP is a filter +in graph domain [31], [34], [35]. GSP usually defines graph +signal models using adjacency matrices as shifts. Signals of a +graph are normally defined at vertices. +Laplacian based GSP originates from spectral graph theory. +High dimensional data are transferred into a low dimensional +space generated by a part of the Laplacian basis [36]. Some +researchers exploited sensor networks [37] to achieve dis- +tributed processing of graph signals. Other researchers solved +the problem globally under the assumption that the graph is +smooth. Unlike adjacency matrix based GSP, Laplacian matrix +is symmetric with real and non-negative edge weights, which +is used to index undirected graphs. +Although the models use different matrices as basic shifts, +most of the notions in GSP are derived from signal processing. +Notions with different definitions in these models may have +similar meanings. All of them correspond to concepts in signal +processing. Signals in GSP are values defined on graphs, and +they are usually written as a vector, s= [s0;s1;:::;sN1]2 +CN: N is the number of vertices, and each element in the +vector represents the value on a vertex. Some studies [26] +allow complex-value signals, even though most applications +are based on real-value signals. +In the context of adjacency matrix based GSP, a graph can +be represented as a triple G(V;E;W), whereVis the vertex +set,Eis the edge set and Wis the adjacency matrix. With +the definition of graphs, we can also define degree matrix +Dii=di, whereDis a diagonal matrix, and diis the degree +of vertexi. Graph Laplacian is defined as L=DW, and +normalized Laplacian is defined as Lnorm =D1=2LD1=2. +Filters in signal processing can be seen as a function that +amplifies or reduces relevant frequencies, eliminating irrele- +vant ones. Matrix multiplication in linear space equals to scale +changing, which is identical with filter operation in frequency +domain. It is obvious that we can use matrix multiplication as +a filter in GSP, which is written as sout=Hsin, whereH +stands for a filter. +Shift is an important concept to describe variation in sig- +nal, and time-invariant signals are used frequently [31]. In +fact, there are different choices of shifts in GSP. Adjacency +matrix based GSP uses Aas shift. Laplacian based GSP uses +L[32], and some researchers also use other matrices [38]. +By following time invariance in traditional signal processing, +shift invariance is defined in GSP. If filters are commutative +with shift, they are shift-invariant, which can be written asIEEE TRANSACTIONS ON ARTIFICIAL INTELLIGENCE, VOL. 00, NO. 0, 2021 5 +AH =HA . It is proved that shift-invariant filter can be +represented by the shift. The properties of shift are vital, and +they determine the fashion of other definitions such as Fourier +transform and frequency. +In adjacency matrix based GSP, eigenvalue decomposition +of shiftAisA=VV1.Vis the matrix of eigenvectors +[v0;v1;:::;vN1]and +=2 +640 +... +N13 +75 +is a diagonal matrix of eigenvalues. The Fourier transform +matrix is the inverse of V, i.e.,F=V1. Frequency of shift +is defined as total variation, which states the difference after +shift +TVG=jjvk1 +maxAvkjj1; +where1 +maxis a normalized factor of matrix. It means that the +frequencies of eigenvalue far away from the largest eigenval- +ues on complex plane are large. A large frequency means that +signals are changed with a large scale after shift filtering. The +differences between minimum and maximum can be seen in +Fig. 4. Generally, the total variation tends to be relatively low +with larger frequency, and vice versa. Eigenvectors of larger +eigenvalues can be used to construct low-frequency filters, +which capture fundamental characteristics, and smaller ones +can be employed to capture the variation among neighbor +nodes. +For topology learning problems, we can distinguish the +corresponding solutions depending on known information. +When topology information is partly known, we can use the +known information to infer the whole graph. In case the +topology information is unknown while we still can observe +the signals on the graph, the topology structure has to be +inferred from the signals. The former one is often solved as a +sampling and recovery problem, and blind topology inference +is also known as graph topology (or structure) learning. +2) Sampling and Recovery: Sampling is not a new concept +defined in GSP. In conventional signal processing, we normally +need to reconstruct original signals with the least samples +and retain all information of original signals for a sampling +problem. Few samples lead to the lack of information and more +samples need more space to store. The well-known Nyquist- +Shannon sampling theorem gives the sufficient condition of +perfect recovery of signals in time domain. +Researchers have migrated the sampling theories into GSP +to study the sampling problem on graphs. As the volume of +data is large in some real-world applications such as sensor +networks and social networks, sampling less and recover- +ing better are vital for GSP. In fact, most algorithms and +frameworks solving sampling problems require that the graph +models correlations within signals observed on it [39]. The +sampling problem can be defined as reconstructing signals +from samples on a subset of vertices, and signals in it are +usually band-limited. Nyquist-Shannon sampling theorem was +extended to graph signals in [40]. Based on the normalized +Laplacian matrix, sampling theorem and cut-off frequency are +(a) The maximum frequency +(b) The minimum frequency +Fig. 4: Illustration of difference between minimum and max- +imum frequencies. +defined for GSP. Moreover, the authors provided a method for +computing cut-off frequency from a given sampling set and a +method for choosing sampling set for a given bandwidth. It +should be noted that the sampling theorem proposed therein +is merely applied to undirected graph. As Laplacian matrix +represents undirected graphs only, sampling theory for directed +graph adopts adjacent matrix. An optimal operator with a +guarantee for perfect recovery was proposed in [35], and it +is robust to noise for general graphs. +One of the explicit distinctions between classical signal +processing and GSP is that signals of the former fall in regular +domain while the latter falls in irregular domain. For sampling +and recovery problems, classical signal processing samples +successive signals and can recover successive signals from +samplings. GSP samples a discrete sequence, and recovers the +original sequences from samplings. By following this order, +the solution is generally separated into two parts, i.e., findingIEEE TRANSACTIONS ON ARTIFICIAL INTELLIGENCE, VOL. 00, NO. 0, 2021 6 +sampling vertex sets and reconstructing original signals based +on various models. +When the size of the dataset is small, we can handle the +signal and shift directly. However, for a large-scale dataset, +some algorithms require matrix decomposition to obtain fre- +quencies and save eigenvalues in the procedure, which are +almost impossible to realize. As a simple technique applicable +to large-scale datasets, a random method can also be used in +sampling. Puy et al. [41] proposed two sample strategies: a +non-adaptive one depending on a parameter and an adaptive +random sampling strategy. By relaxing the optimized con- +straint, they extended random sampling to large scale graphs. +Another common strategy is greedy sampling. For example, +Shomorony and Avestimehr [42] proposed an efficient method +based on linear algebraic conditions that can exactly compute +cut-off frequency. Chamon and Ribeiro [43] provided near- +optimal guarantee for greedy sampling, which guarantees the +performance of greedy sampling in the worst cases. +All of the sampling strategies mentioned above can be +categorized as selecting sampling, where signals are observed +on a subset of vertices [43]. Besides selecting sampling, there +exists a type of sampling called aggregation sampling [44], +which uses observations taken at a single vertex as input, +containing a sequential applications of graph shift operator. +Similar to classical signal processing, the reconstruction +task on graphs can also be interpreted as data interpolation +problem [45]. By projecting the samples on a proper signal +space, researchers obtain interpolated signals. Least squares +reconstruction is an available method in practice. Gadde and +Ortega [46] defined a generative model for signal recovery +derived from a pairwise Gaussian random field (GRF) and +a covariance matrix on graphs. Under sampling theorem, the +reconstruction of graph signals can be viewed as the maximum +posterior inference of GRF with low-rank approximation. +Wang et al. [47] aimed at achieving the distributed reconstruc- +tion of time-varying band limited signal, where the distributed +least squares reconstruction (DLSR) was proposed to recover +the signals iteratively. DLSR can track time-varying signals +and achieve perfect reconstruction. Di Lorenzo et al. [48] +proposed a linear mean squares (LMS) strategy for adaptive +estimation. LMS enables online reconstruction and tracking +from the observation on a subset of vertices. It also allows the +subset to vary over time. Moreover, a sparse online estimation +was proposed to solve the problems with unknown bandwidth. +Another common technique for recovering original signals +is smoothness. Smoothness is used for inferring missing values +in graph signals with low frequencies. Wang et al. [17] +defined the concept of local set. Based on the definition +of graph signals, two iterative methods were proposed to +recover band limited signals on graphs. Besides, Romero +et al. [49] advocated kernel regression as a framework for +GSP modeling and reconstruction. For parameter selection +in estimators, two multi-kernel methods were proposed to +solve a single optimization problem as well. In addition, +some researchers investigated different recovery problems with +compressed sensing [50]. +In addition, there exists some researches on sampling of +different kinds of signals such as smooth graph signals, piece-wise constant signals and piece-wise smooth signals [51]. +Chen et al. [51] gave a uniform framework to analyze graph +signals. The reconstruction of a known graph signal was stud- +ied in [52], where the signal is sparse, which means only a few +vertices are non-zeros. Three kinds of reconstruction schemes +corresponding to various seeding patterns were examined. +By analyzing single simultaneous injection, single successive +value injection, and multiple successive simultaneous injec- +tions, the conditions for perfect reconstruction on any vertices +were derived. +3) Learning Topology Structure from Data: In most appli- +cation scenes, graphs are constructed according to connections +of entity correlations. For example, in sensor networks, the cor- +relations between sensors are often consistent with geographic +distance. Edges in social networks are defined as relations such +as friends or colleagues [53]. In biochemical networks, edges +are generated by interactions. Although GSP is an efficient +framework for solving problems on graphs such as sampling, +reconstruction, and detection, there lacks a step to extract +relations from datasets. Connections exist in many datasets +without explicit records. Fortunately, they can be inferred in +many ways. +As a result, researchers want to learn complete graphs from +datasets. The problem of learning graph from a dataset is +stated as estimating graph Laplacian, or graph topology [54]. +Generally, they require the graph to satisfy some properties, +such as sparsity and smoothness. Smoothness is a widespread +assumption in networks generated from datasets. Therefore, it +is usually used to constrain observed signals and provide a +rational guarantee for graph signals. Researchers have applied +it to graph topology learning. The intuition behind smoothness +based algorithms is that most signals on graph are stationary, +and the result filtered by shift tends to be the lowest frequency. +Dong et al. [55] adopted a factor analysis model for graph +signals, and also imposed a Gaussian prior on latent variables +to obtain a Principal Component Analysis (PCA) like represen- +tation. Kalofolias [56] formulated the objective as a weighted +l1problem and designed a general framework to solve it. +Gauss Markov Random Field (GMRF) is also a widely +used theory for graph topology learning in GSP [54], [57], +[58]. The models of GRMF based graph topology learning +select graphs that are more likely to generate signals which are +similar to the ones generated by GMRF. Egilmez et al. [54] +formulated the problem as a maximum posterior parameter +estimation of GMRF, and the graph Laplacian is a precision +matrix. Pavez and Ortega [57] also formulated the problem as +a precision matrix estimation, and the rows and columns are +updated iteratively by optimizing a quadratic problem. Both +of them restrict the result matrix, which should be Laplacian. +In [58], Pavez et al. chose a two steps framework to find +the structure of the underlying graph. First, a graph topology +inference step is employed to select a proper topology. Then, +a generalized graph Laplacian is estimated. An error bound of +Laplacian estimation is computed. In the next step, the error +bound can be utilized to obtain a matrix in a specific form as +the precision matrix estimation. It is one of the first work that +suggests adjusting the model to obtain a graph satisfying the +requirement of various problems.IEEE TRANSACTIONS ON ARTIFICIAL INTELLIGENCE, VOL. 00, NO. 0, 2021 7 +Diffusion is also a relevant model that can be exploited +to solve the topology interfering problem [39], [59]–[61]. +Diffusion refers to that the node continuously influences its +neighborhoods. In graphs, nodes with larger values will have +higher influence on their neighborhood nodes. Using a few +components to represent signals will help to find the main +factors of signal formation. The models of diffusion are often +under the assumption of independent identically-distributed +signals. Pasdeloup et al. [59] gave the concept of valid graphs +to explain signals and assumed that the signals are observed +after diffusion. Segarra et al. [60] agreed that there exists a +diffusion process in the shift, and the signals can be observed. +The signals in [61] were explained as a linear combination of +a few components. +For time series recorded in data, researchers tried to +construct time-sequential networks. For instance, Mei and +Moura [62] proposed a methodology to estimate graphs, which +considers both time and space dependencies and models them +by auto-regressive process. Segarra et al. [63] proposed a +method that can be seen as an extension of graph learning. +The aim of the paper was to solve the problem of joint +identification of a graph filter and its input signal. +For recovery methods, a well-known partial inference prob- +lem is recommendation [45], [64], [65]. The typical algorithm +used in recommendation is collaborative filtering (CF) [66]. +Given the observed ratings in a matrix, the objective of CF is to +estimate the full rating matrix. Huang et al. [65] demonstrated +that collaborative filtering could be viewed as a specific band- +stop graph filter on networks representing correlations between +users and items. Furthermore, linear latent factor methods can +also be modeled as band limited interpretation problem. +4) Discussion: GSP algorithms have strict limitations on +experimental data, thus leading to less real-world applications. +Moreover, GSP algorithms require the input data to be exactly +the whole graph, which means that part of graph data cannot +be the input. Therefore, the computational complexity of this +kind of methods could be significantly high. In comparison +with other kinds of graph learning methods, the scalability of +GSP algorithms is relatively poor. +B. Matrix Factorization Based Methods +Matrix factorization is a method of simplifying a matrix into +its components. These components have a lower dimension +and could be used to represent the original information of +a network, such as relationships among nodes. Matrix fac- +torization based graph learning methods adopt a matrix to +represent graph characteristics like vertex pairwise similarity, +and the vertex embedding can be achieved by factorizing this +matrix [67]. Early graph learning approaches usually utilized +matrix factorization based methods to solve the graph embed- +ding problem. The input of matrix factorization is the non- +relational high dimensional data feature represented as a graph. +The output of matrix factorization is a set of vertex embedding. +If the input data lies in a low dimensional manifold, the graph +learning for embedding can be treated as a dimension-reduced +problem that preserves the structure information. There are +mainly two types of matrix factorization based graph learning.One is graph Laplacian matrix factorization, and the other is +vertex proximity matrix factorization. +1) Graph Laplacian Matrix Factorization: The preserved +graph characteristics can be expressed as pairwise vertex +similarities. Generally, there are two kinds of graph Laplacian +matrix factorization, i.e., transductive and inductive matrix +factorization. The former only embeds the vertices contained +in the training set, and the latter can embed the vertices that are +not contained in the training set. The general framework has +been designed in [68], and the graph Laplacian matrix factor- +ization based graph learning methods have been summarized +in [69]. The Euclidean distance between two feature vectors is +directly adopted in the initial Metric Multidimensional Scaling +(MDS) [70] to find the optimal embedding. The neighborhoods +of vertices are not considered in the MDS, i.e., any pair +of training instances are considered as connected. The data +feature is extracted by constructing a knearest neighbor graph, +and the subsequent studies [67], [71]–[73] tackle this issue. +The topksimilar neighbors of each vertex are connected with +itself. A similar matrix is calculated by exploiting different +methods, and thus the graph characteristics can be preserved +as much as possible. +Recently, researchers have designed more sophisticated +models. The performance of earlier matrix factorization model +Locality Preserving Projection (LPP) can be improved by +introducing an anchor taking advantage of Anchorgraph-based +Locality Preserving Projection (AgLPP) [74], [75]. The graph +structure can be captured by using a local regression model +and a global regression process based on Local and Global +Regressive Mapping (LGRM) [76]. The global geometry can +be preserved by using local spline regression [77]. +More information can be preserved by exploiting the auxil- +iary information. An adjacency graph and a labelled graph +were constructed in [78]. The objective function of LPP +preserves the local geometric structure of the datasets [67]. +An adjacency graph and a relational feedback graph were con- +structed in [79] as well. The graph Laplacian regularization, +k-means and PCA were considered in RF-Semi-NMF-PCA si- +multaneously [80]. Other works, e.g., [81], adopt semi-definite +programming to learn the adjacency graph that maximizes the +pairwise distances. +2) Vertex Proximity Matrix Factorization: Apart from solv- +ing the above generalized eigenvalue problem, another ap- +proach of matrix factorization is to factorize vertex proximity +matrix directly. In general, matrix factorization can be used +to learn the graph structure from non-relational data, and it is +applicable to learn homogeneous graphs. +Based on matrix factorization, vertex proximity can be +approximated in a low dimensional space. The objective of +preserving vertex proximity is to minimize the error. The +Singular Value Decomposition (SVD) of vertex proximity +matrix was adopted in [82]. There are some other approaches +such as regularized Gaussian matrix factorization [83], low- +rank matrix factorization [84], for solving SVD. +3) Discussion: Matrix factorization algorithms operate on +an interaction matrix to decompose several lower dimension +matrices. The process brings some drawbacks. For example, +the algorithms require a large memory when the decomposedIEEE TRANSACTIONS ON ARTIFICIAL INTELLIGENCE, VOL. 00, NO. 0, 2021 8 +matrices become large. In addition, matrix factorization al- +gorithms are not applicable to supervised or semi-supervised +tasks with the training process. +C. Random Walk Based Methods +Random walk is a convenient and effective way to sample +networks [85], [86]. This method can generate sequences +of nodes meanwhile preserving original relations between +nodes. Based on network structure, NRL can generate feature +vectors of vertices so that downstream tasks can mine network +information in a low dimensional space. An example of NRL +is shown in Fig. 5. The image in Euclidean space is shown +in Fig. 5(a), and the corresponding graph in non-Euclidean +space is shown in Fig. 5(b). As one of the most successful +NRL algorithms, random walks play an important role in +dimensionality reduction. +(a) Image in Euclidean space +(b) Graph in non-Euclidean space +Fig. 5: An example of NRL mapping an image from Euclidean +space into non-Euclidean space. +1) Structure Based Random Walks: Graph-structured data +have various data types and structures. The information en- +coded in a graph is related to graph structure and vertex +attributes, which are the two key factors affecting the reason- +ing of networks. In real-world applications, many networks +only have structural information, but lack vertex attribute +information. How to identify network structure information +effectively, such as important vertices and invisible links,attracts the interest of network scientists [87]. Graph data +have high dimensional characteristics. Traditional network +analysis methods cannot be used for analyzing graph data in +a continuous space. +In recent years, various NRL methods have been proposed, +which preserve rich structural information of networks. Deep- +Walk [88] and Node2vec [7] are two representative methods +for generating network representation of basic network topol- +ogy information. These methods use random walk models +to generate random sequences on networks. By treating the +vertices as words and the generated random sequences of +vertices as word sequences (sentences), the models can learn +the embedding representation of the vertices by inputting these +sequences into the Word2vec model [89]–[91]. The principle +of the learning model is to maximize the co-occurrence prob- +ability of vertices such as Word2vec. In addition, Node2vec +shows that network has complex structural characteristics, +and different network structure samplings can obtain different +results. The sampling mode of DeepWalk is not enough +to capture the diversity of connection patterns in networks. +Node2vec designs a random walk sampling strategy, which +can sample the networks with the preference of breadth-first +sampling and depth-first sampling by adjusting the parameters. +The NRL algorithms mentioned above focused on the first- +order proximity information of vertices. Tang et al. [92] +proposed a method called LINE for large-scale network +embedding. LINE can maintain the first and second order +approximations. The first-order neighbor refers to the one- +hop neighbor between two vertices, and the second-order +neighbor is the neighbor with two hops. LINE is not a deep +learning based model, but it is often compared with these edge +modeling based methods. +It has been proved that the network structure information +plays an important role in various network analysis tasks. In +addition to this structural information, network attributes in +the original network space are also critical in modeling the +formation and evolution of the network [93]. +2) Structure and Vertex Information Based Random Walks: +In addition to network topology, many types of networks also +have rich vertex information, such as vertex content or label +in networks. Yang et al. [84] proposed an algorithm called +TADW. The model is based on DeepWalk and considers the +text information of vertices. The MMDW [94] is another +model based on DeepWalk, which is a kind of semi-supervised +network embedding algorithm, by leveraging labelling infor- +mation of vertices to enhance the performance. +Focusing on the structural identity of nodes, Ribeiro et +al. [95] formulated a framework named Struc2vec. The frame- +work considers nodes with similar local structure rather than +neighborhood and labels of nodes. With hierarchy to evaluate +structural similarity, the framework constrains structural sim- +ilarity more stringently. The experiments indicate that Deep- +Walk and Node2vec are worse than Struc2vec which considers +structural identity. There are some other NRL models, such +as Planetoid [96], which learn network representation using +the feature of network structure and vertex attribute informa- +tion. It is well known that vertex attributes provide effective +information for improving network representation and helpIEEE TRANSACTIONS ON ARTIFICIAL INTELLIGENCE, VOL. 00, NO. 0, 2021 9 +to learn embedded vector space. In the case of relatively +sparse network topology, vertex attribute information can be +used as supplementary information to improve the accuracy +of representation. In practice, how to use vertex information +effectively and how to apply this information to network vertex +embedding are the main challenges in NRL. +Researchers not only investigate random walk based NRL +on vertices but also on graphs. Adhikari et al. [97] proposed an +unsupervised scalable algorithm, Sub2Vec, to learn arbitrary +subgraph. To be more specific, they proposed a method to +measure the similarities between subgraphs without disturbing +local proximity. Narayanan et al. [98] proposed graph2vec, +which is a neural embedding framework. Modeling on neural +document embedding models, graph2vec takes a graph as a +document and the subgraph around words as vertices. By +migrating the model to graphs, the performance of graph2vec +significantly exceeds other substructure representation learning +algorithms. +Generally, random walk can be regarded as a Markov +process. The next state of the process is only related to last +state, which is known as Markov chain. Inspired by vertex- +reinforced random walks, Benson et al. [99] presented spacey +random walk, a non-Markovian stochastic process. As a spe- +cific type of a more general class of vertex-reinforced random +walks, it takes the view that the probability of time remained +on each vertex relates to the long term behavior of dynamical +systems. They proved that dynamical systems can converge to +a stationary distribution under sufficient conditions. +Recently, with the development of Generative Adversarial +Network (GAN), researchers combined random walks with +the GAN method [100], [101]. Existing research on NRL can +be divided into generative models and discriminative models. +GraphGAN [100] integrated these two kinds of models and +played a game-theoretical minimax game. With the process +of the game, the performance of the two models can be +strengthened. Random walk is used as a generator in the +game. NetGAN [101] is a generative model that can model +network in real applications. The method takes the distribution +of biased random walk as input, and can produce graphs with +known patterns. It preserves important topology properties and +does not need to define them in model definition. +3) Random Walks in Heterogeneous Networks: In reality, +most networks contain more than one type of vertex, and +hence networks are heterogeneous. Different from homoge- +neous NRL, heterogenous NRL should well reserve various +relationships among different vertices [102]. Considering the +ubiquitous existence of heterogeneous networks, many ef- +forts have been made to learn network representations of +heterogeneous networks. Compared to homogeneous NRL, the +proximity among entities in heterogeneous NRL is more than +a simple measure of distance or closeness. The semantics +among vertices and links should be considered. Some typical +scenarios include knowledge graphs and social networks. +Knowledge graph is a popular research domain in recent +years. A vital part in knowledge base population is relational +inference. The central problem of relational inference is infer- +ring unknown knowledge from the existing facts in knowledge +bases [103]. There are three types of common relationalinference method in general: statistical relational learning +(SRL), latent factor models (LFM) and random walk models +(RWM). Relational learning methods based on statistics lack +generality and scalability. As a result, latent factor model based +graph embedding and relational paths based random walk have +been adopted more widely. +In a knowledge graph, there exist various vertices and +various types of relationships among different vertices. For +example, in a scholar related knowledge graph [2], [28], the +types of vertices include scholar, paper, publication venue, +institution, etc. The types of relationships include coauthor, +citation, publication, etc. The key idea of knowledge graph +embedding is to embed vertices and their relationships into a +low dimensional vector space, while the inherent structure of +the knowledge graph can be reserved [104]. +For relational paths based random walk, the path ranking +algorithm (PRA) is a path finding method using random walks +to generate relational features on graph data [105]. Random +walks in PRA are with restart, and combine features with a +logistic regression. However, PRA cannot predict connection +between two vertices if there does not exist a path between +them. Gardner et al. [106], [107] introduced two ways to +improve the performance of PRA. One method enables more +efficient processing to incorporate new corpus into knowledge +base, while the other method uses vector space to reduce +the sparsity of surface forms. To resolve cascade errors in +knowledge construction, Wang and Cohen [108] proposed a +joint information extraction and knowledge base based model +with a recursive random walk. Using latent context of the text, +the model obtains additional improvement. Liu et al. [109] +developed a new random walk based learning algorithm named +Hierarchical Random-walk inference (HiRi). It is a two-tier +scheme: the upper tier recognizes relational sequence pattern, +and the lower tier captures information from subgraphs of +knowledge bases. +Another widely-investigated type of heterogeneous net- +works is social networks, such as online social networks and +location based social networks. Social networks are heteroge- +neous in nature because of the different types of vertices and +relations. There are two main ways to embed heterogeneous +social networks, including meta path-based approaches and +random walk-based approaches. +A meta path in heterogeneous networks is defined as a +sequence of vertex types encoding significant composite re- +lations among various types of vertices. Aiming to employ +the rich information in social networks by exploiting various +types of relationships among vertices, Fu et al. [110] proposed +HIN2Vec, which is a representation learning framework based +on meta-paths. HIN2Vec is a neural network model and the +meta-paths are well embedded based on two independent +phases, i.e., training data preparation and representation learn- +ing. Experimental results on various social network datasets +show that HIN2Vec model is able to automatically learn vertex +vector in heterogeneous networks to support a variety of +applications. Metapath2vec [111] was designed by formalizing +meta-path based random walks to construct the neighborhoods +of a vertex in heterogeneous networks. It takes the advantage +of a heterogeneous skip-gram model to perform vertex em-IEEE TRANSACTIONS ON ARTIFICIAL INTELLIGENCE, VOL. 00, NO. 0, 2021 10 +bedding. +Meta path based methods require either prior knowledge for +optimal meta-path selection or extended computations for path +length selection. To overcome these challenges, random walk +based approaches have been proposed. Hussein et al. [112] +proposed the JUST model, which is a heterogeneous graph +embedding approach using random walks with jump and stay +strategies so that the aforementioned bias can be overcomed ef- +fectively. Another method which does not require prior knowl- +edge for meta-path definition is MPDRL [113], meta-path +discovery with reinforcement earning. This method employs +the reinforcement learning algorithm to perform multi-hop rea- +soning to generate path instances and then further summarizes +the important meta-paths using the Lowest Common Ancestor +principle. Shi et al. [114] proposed the HERec model, which +utilizes the heterogeneous information network embedding +for providing accurate recommendations in social networks. +HERec is designed based on a random walk based approach +for generating meaningful vertex sequences for heterogeneous +network embedding. HERec can effectively adopt the auxiliary +information in heterogeneous information networks. Other +typical heterogeneous social network embedding approaches +include, e.g., PTE [115] and SHNE [116]. +4) Random Walks in Time-varying Networks: Network is +evolving over time, which means that new vertices may emerge +and new relations may appear. Therefore, it is significant +to capture the temporal behaviour of networks in network +analysis. Many efforts have been made to learn time-varying +network embedding (e.g., dynamic networks or temporal net- +works) [117]. In contrast to static network embedding, time- +varying NRL should consider the network dynamics, which +means that old relationships may become invalid and new links +may appear. +The key of time-varying NRL is to find a suitable way to +incorporate the time characteristic into embedding via reason- +able updating approaches. Nguyen et al. [118] proposed the +CTDNE model for continuous dynamic network embedding +based on random walk with ”chronological” paths which can +only move forward as time goes on. Their model is more +suitable for time-dependent network representation that can +capture the important temporal characteristics of continuous- +time dynamic networks. Results on various datasets show +that CTDNE outperforms static NRL approaches. Zuo et +al. [119] proposed the HTNE model which is a temporal +NRL approach based on the Hawkes process. HTNE can well +integrate the Hawkes process into network embedding so that +the influence of historical neighbors on the current neighbors +can be accurately captured. +For unseen vertices in a dynamical network, Graph- +SAGE [120] was presented to efficiently generate embed- +dings for new vertices in network. In contrast to methods +that training embedding for every vertex in the network, +GraphSAGE designs a function to generate embedding for +a vertex with features of the neighborhoods locally. After +sampling neighbors of a vertex, GraphSAGE uses different +aggregators to update the embedding of the vertex. However, +current graph neural methods are proficient of only learning +local neighborhood information and cannot directly explorethe higher-order proximity and the community structure of +graphs. +5) Discussion: As mentioned before, random walk is a +fundamental way to sample networks. The sequences of nodes +could preserve the information of network structure. However, +there are some disadvantages of this method. For example, +random walk relies on random strategies, which creates some +uncertain relations of nodes. To reduce this uncertainty, it +needs to increase the number of samples, which will signifi- +cantly increase the complexity of algorithms. Some random +walk variants could preserve local and global information +of networks, but they might not be effective in adjusting +parameters to adapt to different types of networks. +D. Deep Learning on Graphs +Deep learning is one of the hottest areas over the past few +years. Nevertheless, it is an attractive and challenging task to +extend the existing neural network models, such as Recurrent +Neural Networks (RNNs) or Convolutional Neural Networks +(CNNs), to graph data. Gori et al. [121] proposed a GNN +model based on recursive neural network. In this model, a +transfer function is implemented, which maps the graph or its +vertices to an m-dimensional Euclidean space. In recent years, +lots of GNN models have been proposed. +1) Graph Convolutional Networks: GCN works on the ba- +sis of grid structure domain and graph structure domain [122]. +Time Domain and Spectral Methods . Convolution is one +of a common operation in deep learning. However, since graph +lacks a grid structure, standard convolution over images or +text cannot be directly applied to graphs. Bruna et al. [122] +extended the CNN algorithm from image processing to the +graph using the graph Laplacian matrix, dubbed as spectral +graph CNN. The main idea is similar to Fourier basis for +signal processing. Based on [122], Henaff et al. [123] defined +kernels to reduced the learning parameters by analogizing +the local connection of CNNs on the image. Defferrard et +al. [124] provided two ways for generalizing CNNs to graph +structure data based on graph theory. One method is to reduce +the parameters by using polynomial kernel, and this method +can be accelerated by using Chebyshev polynomial approx- +imation. The other method is the special pooling method, +which is pooling on the binary tree constructed from vertices. +An improved version of [124] was introduced by Kipf and +Welling [125]. The proposed method is a semi-supervised +learning method for graphs. The algorithm employs an excel- +lent and straightforward neural network followed by a layer- +by-layer propagation rule, which is based on the first-order +approximation of spectral convolution on the graph and can +be directly acted on the graph. +There are some other time domain based methods. Based +on the mixture model of CNNs, for instance, Monti et +al. [126] generalized the CNN to non-Euclidean space. Zhou +and Li [127] proposed a new CNN graph modeling framework, +which designs two modules for graph structure data: K- +order convolution operator and adaptive filtering module. In +addition, the high-order adaptive graph convolution network +(HA-GCN) framework proposed in [127] is a general ar- +chitecture that is suitable for many applications of verticesIEEE TRANSACTIONS ON ARTIFICIAL INTELLIGENCE, VOL. 00, NO. 0, 2021 11 +and graph centers. Manessi et al. [128] proposed a dynamic +graph convolution network algorithm for dynamic graphs. +The core idea of the algorithm is to combine the expansion +of graph convolution with the improved Long Short Term- +Memory networks (LSTM) algorithm, and then train and learn +the downstream recursive unit by using graph structure data +and vertex features. The spectral based NRL methods have +many applications, such as vertex classification [125], traffic +forecasting [129], [130], and action recognition [131]. +Space Domain and Spatial Methods . Spectral graph +theory provides a convolution method on graphs, but many +NRL methods directly use convolution operation on graphs +in space domain. Niepert et al. [132] applied graph labeling +procedures such as Weisfeiler-Lehman kernel on graphs to +generate unique order of vertices. The generated sub-graphs +can be fed to the traditional CNN operation in space domain. +Duvenaud et al. [133] designed Neural fingerprints (FP), which +is a spatial method using the first-order neighbors similar to +the GCN algorithm. Atwood and Towsley [134] proposed an- +other convolution method, called diffusion-convolutional neu- +ral network, which incorporates transfer probability matrix and +replaces the characteristic basis of convolution with diffusion +basis. Gilmer et al. [135] reformulated existing models into +a single common framework, and exploited this framework to +discover new variations. Allamanis et al. [136] represented the +structure of code from syntactic and semantic, and utilized the +GNN method to recognize program structures. +Zhuang and Ma [137] designed dual graph convolution +networks (DGCN), which use diffusion basis and adjacency +basis. DGCN uses two convolutions: one is the characteristic +form of polynomial filter, and the other is to replace the +adjacency matrix with the PPMI (Positive Pointwise Mutual +Information) of the transition probability [89]. Dai et al. [138] +proposed the SSE algorithm, which uses asynchronous ran- +dom to learn vertex representation so as to improve learning +efficiency. In this model, a recursive method is adopted to +learn vertex latent representation and the sampled batch data +are utilized to update parameters. The recursive function of +SSE is calculated from the weighted average of historical state +and new state. Zhu et al. [139] proposed a graph smoothing +splines neural network which exploits non-smoothing node +features and global topological knowledge such as centrality +for graph classification. Gao et al. [140] proposed a large scale +graph convolution network (LGCN) based on vertex feature +information. In order to adapt to the scene of large scale +graphs, they proposed a sub-graph training strategy, which first +trained the sampled sub-graph in a small batch. Based on a +deep generative graph model, a novel method called DeepNC +for inferring the missing parts of a network was proposed +in [141]. +A brief history of deep learning on graphs is shown in Fig. 6. +GNN has attracted lots of attention since 2015, and it is widely +studied and used in various fields. +2) Graph Attention Networks: In sequence-based tasks, +attention mechanism has been regarded as a standard [142]. +GNNs achieve lots of benefits from the expanded model +capacity of attention mechanisms. GATs are a kind of spatial- +based GCNs [143]. It takes the attention mechanism into con-sideration when determining the weights of vertex’s neighbors. +Likewise, Gated Attention Networks (GAANs) also introduced +the multi-head attention mechanism for updating the hidden +state of some vertices [144]. Unlike GATs, GAANs employ a +self-attention mechanism which can compute different weights +for different heads. Some other models such as graph at- +tention model (GAM) were proposed for solving different +problems [145]. Take GAM as an example, the purpose of +GAM is to handle graph classification. Therefore, GAM is set +to process informative parts by visiting a sequence of signifi- +cant vertices adaptively. The model of GAM contains LSTM +network, and some parameters contain historical information, +policies, and other information generated from exploration of +the graph. Attention Walks (AWs) are another kind of learning +model based on GNN and random walks [146]. In contrast +to DeepWalk, AWs use differentiable attention weights when +factorizing the co-occurrence matrix [88]. +3) Graph Auto-Encoders: GAE uses GNN structure to +embed network vertices into low dimensional vectors. One +of the most general solutions is to employ a multi-layer +perception as the encoder for inputs [147]. Therein the decoder +reconstructs neighborhood statistics of the vertex. PPMI or +the first and the second nearest neighborhood can be taken +into statistics [148], [149]. Deep neural networks for graph +representations (DNGR) employ PPMI. Structural deep net- +work embedding (SDNE) employs stacked auto-encoder to +maintain both the first-order and the second-order proximity. +Auto-encoder [150] is a traditional deep learning model, +which can be classified as a self-supervised model [151]. +Deep recursive network embedding (DRNE) reconstructs some +vertices’ hidden state rather than the entire graph [152]. It has +been found that if we regard GCN as an encoder, and combine +GCN with GAN or LSTM with GAN, then we can design +the auto-encoder for graphs. Generally speaking, DNGR and +SDNE embed vertices by the given structure features, while +other methods such as DRNE learn both topology structure +and content features [148], [149]. Variational graph auto- +encoder [153] is another successful approach that employs +GCN as an encoder and a link prediction layer as a decoder. +Its successor, adversarially regularized variational graph auto- +encoder [154], adds a regularization process with an adversar- +ial training approach to learn a more robust embedding. +4) Graph Generative Networks: The purpose of graph +generative networks is to generate graphs according to the +given observed set of graphs. Many previous methods of graph +generative networks have their own application domains. For +example, in natural language processing, the semantic graph +or the knowledge graph is generated based on the given +sentences. Some general methods have been proposed recently. +One kind of them considers the generation process as the for- +mation of vertices and edges. Another kind is to employ gener- +ative adversarial training. Some GCNs based graph generative +networks such as molecular generative adversarial networks +(MolGAN) integrate GNN with reinforcement learning [155]. +Deep generative models of graphs (DGMG) achieves a hidden +representation of existing graphs by utilizing spatial-based +GCNs [156]. There are some knowledge graph embedding +algorithms based on GAN and Zero-Shot Learning [157]. VyasIEEE TRANSACTIONS ON ARTIFICIAL INTELLIGENCE, VOL. 00, NO. 0, 2021 12 +Fig. 6: A brief history of algorithms of deep learning on graphs. +et al. [158] proposed a Generalized Zero-Shot learning model, +which can find unseen semantic in knowledge graphs. +5) Graph Spatial-Temporal Networks: Graph spatial- +temporal networks simultaneously capture the spatial and tem- +poral dependence of graphs. The global structure is included in +the spatial-temporal graphs, and the input of each vertex varies +with the change of time. For example, in traffic networks, each +sensor records the traffic speed of a road continuously as a +vertex, in which the edge of the traffic networks is determined +by the distance between the sensor pairs [129]. The goal of a +spatial-temporal network can be to predict future vertex values +or labels, or to predict spatial-temporal graph labels. Recent +studies in this direction have discussed the use of GCNs, +the combination of GCNs with RNN or CNN, and recursive +structures for graph structures [130], [131], [159]. +6) Discussion: In this context, the task of graph learning +can be seen as optimizing the objective function by using +gradient descent algorithms. Therefore the performance of +deep learning based NRL models is influenced by gradient +descent algorithms. They may encounter challenges like local +optimal solutions and the vanishing gradient problem. +III. A PPLICATIONS +Many problems can be solved by graph learning methods, +including supervised, semi-supervised, unsupervised, and re- +inforcement learning. Some researchers classify the applica- +tions of graph learning into three categories, i.e., structural +scenarios, non-structural scenarios, and other application sce- +narios [18]. Structural scenarios refer to the situation where +data are performed in explicit relational structures, such as +physical systems, molecular structures, and knowledge graphs. +Non-structural scenarios refer to the situation where data are +with unclear relational structures, such as images and texts. +Other application scenarios include, e.g., integrating models +and combinatorial optimization problems. Table II lists the +neural components and applications of various graph learning +methods. +A. Datasets and Open-source Libraries +There are several datasets and benchmarks used to evaluate +the performance of graph learning approaches for various tasks +such as link prediction, node classification, and graph visual- +ization. For instance, datasets like Cora1(citation network), +1https://relational.fit.cvut.cz/dataset/CORAPubmed2(citation network), BlogCatalog3(social network), +Wikipedia4(language network) and PPI5(biological network) +include nodes, edges, labels or attributes of nodes. Some +research institutions developed graph learning libraries, which +include common and classical graph learning algorithms. For +example, OpenKE6is a Python library for knowledge graph +embedding based on PyTorch. The open-source framework has +the implementations of RESCAL, HolE, DistMult, ComplEx, +etc. CogDL7is a graph representation learning framework, +which can be used for node classification, link prediction, +graph classification, etc. +B. Text +Many data are in textual form coming from various re- +sources like web pages, emails, documents (technical and +corporate), books, digital libraries and customer complains, +letters, patents, etc. Textual data are not well structured for +obtaining any meaningful information as text often contains +rich context information. There exist abundant applications +around text, including text classification, sequence labeling, +sentiment classification, etc. Text classification is one of +the most classical problems in natural language processing. +Popular algorithms proposed to handle this problem include +GCNs [120], [125], GATs [143], Text GCNs [160], and +Sentence LSTM [161]. Sentence LSTM has also been applied +to sequence labeling, text generation, multi-hop reading com- +prehension, etc [161]. Syntactic GCN was proposed to solve +semantic role labeling and neural machine translation [162]. +Gated Graph Neural Networks (GGNNs) can also be used to +address neural machine translation and text generation [163]. +For relational extraction, Tree LSTM, graph LSTM, and GCN +are better solutions [164]. +C. Images +Graph learning applications pertaining to images include +social relationship understanding, image classification, visual +question answering, object detection, region classification, and +semantic segmentation, etc. For social relationship understand- +ing, for instance, graph reasoning model (GRM) is widely +2https://catalog.data.gov/dataset/pubmed +3http://networkrepository.com/soc-BlogCatalog.php +4https://en.wikipedia.org/wiki/Wikipedia:Database download +5https://openwetware.org/wiki/Protein-protein interaction databases +6https://github.com/thunlp/OpenKE +7https://github.com/THUDM/cogdl/IEEE TRANSACTIONS ON ARTIFICIAL INTELLIGENCE, VOL. 00, NO. 0, 2021 13 +TABLE II: Summary of graph learning methods and their applications +Categories Algorithms Neural Component Applications +Time Domain and Spectral MethodsSNLCN [122] Graph Neural Network Classification +DCN [123] Spectral Network Classification +ChebNet [124] Convolution Network Classification +GCN [125] Spectral Network Classification +HA-GCN [127] GCN Classification +Dynamic GCN [128] GCN, LSTM Classification +DCRNN [129] Diffusion Convolution Network Traffic Forecasting +ST-GCN [131] GCN Action Recognition +Space Domain and Spatial MethodsPATCHY-SAN [132] Convolutional Network Runtime Analysis, +Feature Visualization, +Graph Classification +Neural FP [133] Sub-graph Classification +DCNN [134] DCNN Classification +DGCN [137] Graph-Structure-Based +Convolution, PPMI-Based +Convolution.Classification +SSE [138] Vertex Classification +LGCN [140] Convolutional Neural Network Vertex Classification +STGCN [130] Gated Sequential Convolution Traffic Forecasting +Deep Learning Model Based MethodsGATs [143] +Attention Neural NetworkClassification +GAAN [144] Vertex Classification +GAM [145] Graph Classification +Aws [146] +Auto-encoder Neural NetworkLink Prediction, +Sensitivity Analysis, +Vertex Classification +SDNE [149] Classification, +Link Prediction, +Visualization +DNGR [148] Clustering, Visualization +DRNE [152] Regular Equivalence Predic- +tion, +Structural Role Classifica- +tion, +Network Visualization +MolGAN [155] +Generative Neural NetworkGenerative Model +DGMG [156] Molecule Generation +DCRNN [129] Diffusion Convolution Network Traffic Forecasting +STGCN [130] Gated Sequential Convolution +ST-GCN [131] GCNs Action Recognition +used [165]. Since social relationships such as friendships +are the basis of social networks in real world, automatically +interpreting these relationships is important for understanding +human behaviors. GRM introduces GGNNs to learn a propa- +gation mechanism. Image classification is a classical problem, +in which GNNs have demonstrated promising performance. +Visual question answering (VQA) is a learning task that in- +volves both computer vision and natural language processing. +A VQA system takes the form of a certain pictures and its +open natural language question as input, in order to generate +a natural language answer as output. Generally speaking, VQA +is question-and-answer for a given picture. GGNNs have been +exploited to help with VQA [166].D. Science +Graph learning has been widely adopted in science. Model- +ing real-world physical systems is one of the most fundamental +perspectives in understanding human intelligence. Represent- +ing objects as vertices and relations as edges between them +is a simple but effective way to perform physics. Battaglia et +al. [167] proposed interaction networks (IN) to predict and +infer abundant physical systems, in which IN takes objects +and relationships as input. Based on IN, the interactions can +be reasoned and the effects can be applied. Therefore, physical +dynamics can be predicted. Visual interaction networks (VIN) +can make predictions from pixels by firstly learning a stateIEEE TRANSACTIONS ON ARTIFICIAL INTELLIGENCE, VOL. 00, NO. 0, 2021 14 +code from two continuous input frames per object [168]. +Other graph networks based models have been developed to +address chemistry and biology problems. Calculating molecu- +lar fingerprints, i.e., using feature vectors to represent molec- +ular, is a central step. Researchers [169] proposed neural +graph fingerprints using GCNs to calculate substructure feature +vectors. Some studies focused on protein interface prediction. +This is a challenging issue with significant applications in +biology. Besides, GNNs can be used in biomedical engineering +as well. Based on protein-protein interaction networks, Rhee et +al. [170] used graph convolution and protein relation networks +to classify breast cancer subtypes. +E. Knowledge Graphs +Various heterogeneous objects and relationships are re- +garded as the basis for a knowledge graph [171]. GNNs can +be applied in knowledge base completion (KBC) for solving +the out-of-knowledge-base (OOKB) entity problem [172]. The +OOKB entities are connected to existing entities. Therefore, +the embedding of OOKB entities can be aggregated from +existing entities. Such kind of algorithms achieve reasonable +performance in both settings of KBC and OOKB. Likewise, +GCNs can also be used to solve the problem of cross-lingual +knowledge graph alignment. The main idea of the model is to +embed entities from different languages into an integrated em- +bedding space. Then the model aligns these entities according +to their embedding similarities. +Generally speaking, knowledge graph embedding can be +categorized into two types: translational distance models and +semantic matching models. Translational distance models aim +to learn the low dimensional vector of entities in a knowledge +graph by employing distance-based scoring functions. These +methods calculate the plausibility as the distance between +two entities after a translation measured by the relationships +between them. Among current translational distance models, +TransE [173] is the most influential one. TransE can model the +relationship of entities by interpreting them as translations op- +erating on the low dimensional embedding. Inspired by TranE, +TranH [174] was proposed to overcome the disadvantages +of TransE in dealing with 1-to-N, N-to-1, and N-to-N rela- +tions by introducing relation-specific hyperplanes. Instead of +hyperplanes, TransR [175] introduces relation-specific spaces +to solve the flows in TransE. Meanwhile, various extensions +of TransE have been proposed to enhance knowledge graph +embeddings, such as TransD [176] and TransF [177]. On the +basis of TransE, DeepPath [178] incorporates reinforcement +learning methods for learning relational paths in knowledge +graphs. By designing a complex reward function involving +accuracy, efficiency and path diversity, the path finding process +is better controlled and more flexible. +Semantic matching models utilize the similarity-based scor- +ing functions. They measure the plausibility among entities +by matching latent semantics of entities and relations in low +dimensional vector space. Typical models of this type include +RESCAL [179], DistMult [180], ANALOGY [181], etc.F . Combinatorial Optimization +Classical problems such as traveling salesman problem +(TSP) and minimum spanning tree (MST) have been solved +by using different heuristic solutions. Recently, deep neural +networks have been applied to these problems. Some solutions +make further use of GNNs thanks to their structures. Bello et +al. [182] first proposed such kind of methods to solve TSP. +Their method mainly contains two steps, i.e., a parameterized +reward pointer network and a strategy gradient module for +training. Khalil et al. [183] improved this work with GNN +and achieved better performance by two main procedures. +First, they used structure2vec to achieve vertex embedding and +then input them into Q-learning module for decision-making. +This work also proves the embedding ability of GNN. Nowak +et al. [184] focused on the secondary assignment problem, +i.e., measuring the similarity of two graphs. The GNN model +learns each graph’s vertex embedding and uses the attention +mechanism to match the two graphs. Other studies use GNNs +directly as the classifiers, which can perform the intensive +prediction on graphs with two sides. The rest of the model +facilitates diverse choices and effective training. +IV. O PEN ISSUES +In this section, we briefly summarize several future research +directions and open issues for graph learning. +Dynamic Graph Learning : For the purpose of graph learn- +ing, most existing algorithms are suitable for static networks +without specific constraints. However, dynamic networks such +as traffic networks vary over time. Therefore, they are hard to +deal with. Dynamic graph learning algorithms have rarely been +studied in the literature. It is of significant importance that +dynamic graph learning algorithms are designed to maintain +good performance, especially in the case of dynamic graphs. +Generative Graph Learning : Inspired by the generative +adversarial networks, generative graph learning algorithms can +unify the generative and discriminative models by playing a +game-theoretical min-max game. This generative graph learn- +ing method can be used for link prediction, network evolution, +and recommendation by boosting the performance of genera- +tive and discriminative models alternately and iteratively. +Fair Graph Learning : Most graph learning algorithms rely +on deep neural networks, and the resulting vectors may have +captured undesired sensitive information. The bias existing +in the network is reinforced, and hence it is of significant +importance to integrate the fair metrics into the graph learning +algorithms to address the inherent bias issue. +Interpretability of Graph Learning : The models of graph +learning are generally complex by incorporating both graph +structure and feature information. The interpretability of graph +learning (based) algorithms remains unsolved since the struc- +tures of graph learning algorithms are still a black box. For +example, drug discovery can be achieved by graph learning al- +gorithms. However, it is unknown how this drug is discovered +as well as the reason behind this discovery. The interpretability +behind graph learning needs to be further studied.IEEE TRANSACTIONS ON ARTIFICIAL INTELLIGENCE, VOL. 00, NO. 0, 2021 15 +V. C ONCLUSION +This survey gives a general description of graph learning, +and provides a comprehensive review of the state-of-the-art +graph learning methods. We examined existing graph learning +methods under four categories: graph signal processing based +methods, matrix factorization based methods, random walk +based methods, and deep learning based methods. The ap- +plications of graph learning methods mainly under these four +categories in areas such as text, images, science, knowledge +graphs, and combinatorial optimization are outlined. We also +discuss some future research directions in the field of graph +learning. Graph learning is currently a hot area which is grow- +ing at an unprecedented speed. We do hope that this survey +will help researchers and practitioners with their research and +development in graph learning and related areas. +ACKNOWLEDGMENTS +The authors would like to thank Prof. Hussein Abbass at +University of New South Wales, Yuchen Sun, Jiaying Liu, +Hao Ren at Dalian University of Technology, and anonymous +reviewers for their valuable comments and suggestions. +REFERENCES +[1] S. Fortunato, C. T. Bergstrom, K. B ¨orner, J. A. Evans, D. Helbing, +S. Milojevi ´c, A. M. Petersen, F. Radicchi, R. Sinatra, B. Uzzi et al. , +“Science of science,” Science , vol. 359, no. 6379, 2018, eaao0185. +[2] J. Liu, J. Ren, W. Zheng, L. Chi, I. Lee, and F. Xia, “Web of scholars: +A scholar knowledge graph,” in the 43rd International ACM SIGIR +Conference on Research and Development in Information Retrieval +(SIGIR) , 2020, pp. 2153–2156. +[3] J. Liu, F. Xia, L. Wang, B. Xu, X. Kong, H. Tong, and I. King, +“Shifu2: A network representation learning based model for advisor- +advisee relationship mining,” IEEE Transactions on Knowledge and +Data Engineering , vol. 33, no. 4, pp. 1763–1777, 2021. +[4] D. Zhang, J. Yin, X. Zhu, and C. Zhang, “Network representation +learning: A survey,” IEEE Transactions on Big Data , vol. 6, no. 1, +pp. 3–28, 2020. +[5] K. Sun, J. Liu, S. Yu, B. Xu, and F. Xia, “Graph force learning,” +inIEEE International Conference on Big Data (BigData) , 2020, pp. +2987–2994. +[6] F. Xia, J. Wang, X. Kong, D. Zhang, and Z. Wang, “Ranking station im- +portance with human mobility patterns using subway network datasets,” +IEEE Transactions on Intelligent Transportation Systems , vol. 21, no. 7, +pp. 2840–2852, 2020. +[7] A. Grover and J. Leskovec, “node2vec: Scalable feature learning for +networks,” in Proceedings of the 22nd ACM SIGKDD International +Conference on Knowledge Discovery and Data Mining . ACM, 2016, +pp. 855–864. +[8] K. Sun, L. Wang, B. Xu, W. Zhao, S. W. Teng, and F. Xia, “Net- +work representation learning: From traditional feature learning to deep +learning,” IEEE Access , vol. 8, no. 1, pp. 205 600–205 617, 2020. +[9] S. Yu, F. Xia, J. Xu, Z. Chen, and I. Lee, “Offer: A motif dimensional +framework for network representation learning,” in The 29th ACM +International Conference on Information and Knowledge Management +(CIKM) , 2020, pp. 3349–3352. +[10] Y . Bengio, A. Courville, and P. Vincent, “Representation learning: A +review and new perspectives,” IEEE Transactions on Pattern Analysis +and Machine Intelligence , vol. 35, no. 8, pp. 1798–1828, 2013. +[11] T. Guo, F. Xia, S. Zhen, X. Bai, D. Zhang, Z. Liu, and J. Tang, +“Graduate employment prediction with bias,” in Thirty-Fourth AAAI +Conference on Artificial Intelligence (AAAI) , 2020, pp. 670–677. +[12] F. Xia, A. M. Ahmed, L. T. Yang, J. Ma, and J. J. Rodrigues, “Ex- +ploiting social relationship to enable efficient replica allocation in ad- +hoc social networks,” IEEE Transactions on Parallel and Distributed +Systems , vol. 25, no. 12, pp. 3167–3176, 2014. +[13] J. Zhang, W. Wang, F. Xia, Y .-R. Lin, and H. Tong, “Data-driven +computational social science: A survey,” Big Data Research , vol. 21, +p. 100145, 2020.[14] F. Xia, A. M. Ahmed, L. T. Yang, and Z. Luo, “Community-based +event dissemination with optimal load balancing,” IEEE Transactions +on Computers , vol. 64, no. 7, pp. 1857–1869, 2014. +[15] F. Xia, H. Wei, S. Yu, D. Zhang, and B. Xu, “A survey of measures +for network motifs,” IEEE Access , vol. 7, no. 1, pp. 106 576–106 587, +2019. +[16] J. Leskovec and C. Faloutsos, “Sampling from large graphs,” in +Proceedings of the 12th ACM SIGKDD International Conference on +Knowledge Discovery and Data Mining . ACM, 2006, pp. 631–636. +[17] X. Wang, P. Liu, and Y . Gu, “Local-set-based graph signal reconstruc- +tion,” IEEE Transactions on Signal Processing , vol. 63, no. 9, pp. +2432–2444, 2015. +[18] Z. Zhang, P. Cui, and W. Zhu, “Deep learning on graphs: A survey,” +IEEE Transactions on Knowledge and Data Engineering , 2020. +[19] J. Xu, S. Yu, K. Sun, J. Ren, I. Lee, S. Pan, and F. Xia, “Multivariate +relations aggregation learning in social networks,” in ACM/IEEE Joint +Conference on Digital Libraries (JCDL) , 2020, pp. 77–86. +[20] H. D. Bedru, S. Yu, X. Xiao, D. Zhang, L. Wan, H. Guo, and F. Xia, +“Big networks: A survey,” Computer Science Review , vol. 37, p. +100247, 2020. +[21] P. Goyal and E. Ferrara, “Graph embedding techniques, applications, +and performance: A survey,” Knowledge-Based Systems , vol. 151, pp. +78–94, 2018. +[22] H. Cai, V . W. Zheng, and K. C.-C. Chang, “A comprehensive survey +of graph embedding: Problems, techniques, and applications,” IEEE +Transactions on Knowledge and Data Engineering , vol. 30, no. 9, pp. +1616–1637, 2018. +[23] M. Nickel, K. Murphy, V . Tresp, and E. Gabrilovich, “A review of +relational machine learning for knowledge graphs,” Proceedings of the +IEEE , vol. 104, no. 1, pp. 11–33, 2016. +[24] L. Akoglu, H. Tong, and D. Koutra, “Graph based anomaly detection +and description: a survey,” Data Mining and Knowledge Discovery , +vol. 29, no. 3, pp. 626–688, 2015. +[25] Z. Wu, S. Pan, F. Chen, G. Long, C. Zhang, and P. S. Yu, “A +comprehensive survey on graph neural networks,” IEEE Transactions +on Neural Networks and Learning Systems , vol. 32, no. 1, pp. 4–24, +2021. +[26] A. Ortega, P. Frossard, J. Kova ˇcevi´c, J. M. Moura, and P. Van- +dergheynst, “Graph signal processing: Overview, challenges, and ap- +plications,” Proceedings of the IEEE , vol. 106, no. 5, pp. 808–828, +2018. +[27] W. Huang, T. A. Bolton, J. D. Medaglia, D. S. Bassett, A. Ribeiro, and +D. Van De Ville, “A graph signal processing perspective on functional +brain imaging,” Proceedings of the IEEE , vol. 106, no. 5, pp. 868–885, +2018. +[28] F. Xia, W. Wang, T. M. Bekele, and H. Liu, “Big scholarly data: A +survey,” IEEE Transactions on Big Data , vol. 3, no. 1, pp. 18–35, 2017. +[29] Y . LeCun, Y . Bengio, and G. Hinton, “Deep learning,” Nature , vol. +521, no. 7553, p. 436, 2015. +[30] J. Liu, X. Kong, F. Xia, X. Bai, L. Wang, Q. Qing, and I. Lee, “Artificial +intelligence in the 21st century,” IEEE Access , vol. 6, pp. 34 403– +34 421, 2018. +[31] A. Sandryhaila and J. M. Moura, “Discrete signal processing on +graphs,” IEEE Transactions on Signal Processing , vol. 61, no. 7, pp. +1644–1656, 2013. +[32] D. Shuman, S. Narang, P. Frossard, A. Ortega, and P. Vandergheynst, +“The emerging field of signal processing on graphs: Extending high- +dimensional data analysis to networks and other irregular domains,” +IEEE Signal Processing Magazine , vol. 3, no. 30, pp. 83–98, 2013. +[33] M. Puschel and J. M. Moura, “Algebraic signal processing theory: +Foundation and 1-d time,” IEEE Transactions on Signal Processing , +vol. 56, no. 8, pp. 3572–3585, 2008. +[34] A. Sandryhaila and J. M. Moura, “Discrete signal processing on graphs: +Graph filters,” in 2013 IEEE International Conference on Acoustics, +Speech and Signal Processing . IEEE, 2013, pp. 6163–6166. +[35] S. Chen, R. Varma, A. Sandryhaila, and J. Kova ˇcevi´c, “Discrete signal +processing on graphs: Sampling theory,” IEEE Transactions on Signal +Processing , vol. 63, no. 24, pp. 6510–6523, 2015. +[36] U. V on Luxburg, “A tutorial on spectral clustering,” Statistics and +Computing , vol. 17, no. 4, pp. 395–416, 2007. +[37] X. Zhu and M. Rabbat, “Graph spectral compressed sensing for +sensor networks,” in 2012 IEEE International Conference on Acoustics, +Speech and Signal Processing (ICASSP) . IEEE, 2012, pp. 2865–2868. +[38] A. Gavili and X.-P. Zhang, “On the shift operator, graph frequency, +and optimal filtering in graph signal processing,” IEEE Transactions +on Signal Processing , vol. 65, no. 23, pp. 6303–6318, 2017.IEEE TRANSACTIONS ON ARTIFICIAL INTELLIGENCE, VOL. 00, NO. 0, 2021 16 +[39] B. Pasdeloup, M. Rabbat, V . Gripon, D. Pastor, and G. Mercier, +“Graph reconstruction from the observation of diffused signals,” in +2015 53rd Annual Allerton Conference on Communication, Control, +and Computing (Allerton) . IEEE, 2015, pp. 1386–1390. +[40] A. Anis, A. Gadde, and A. Ortega, “Towards a sampling theorem for +signals on arbitrary graphs,” in 2014 IEEE International Conference +on Acoustics, Speech and Signal Processing (ICASSP) . IEEE, 2014, +pp. 3864–3868. +[41] G. Puy, N. Tremblay, R. Gribonval, and P. Vandergheynst, “Random +sampling of bandlimited signals on graphs,” Applied and Computa- +tional Harmonic Analysis , vol. 44, no. 2, pp. 446–475, 2018. +[42] H. Shomorony and A. S. Avestimehr, “Sampling large data on graphs,” +in2014 IEEE Global Conference on Signal and Information Processing +(GlobalSIP) . IEEE, 2014, pp. 933–936. +[43] L. F. Chamon and A. Ribeiro, “Greedy sampling of graph signals,” +IEEE Transactions on Signal Processing , vol. 66, no. 1, pp. 34–47, +2018. +[44] A. G. Marques, S. Segarra, G. Leus, and A. Ribeiro, “Sampling of +graph signals with successive local aggregations.” IEEE Transactions +Signal Processing , vol. 64, no. 7, pp. 1832–1843, 2016. +[45] S. K. Narang, A. Gadde, and A. Ortega, “Signal processing techniques +for interpolation in graph structured data,” in 2013 IEEE International +Conference on Acoustics, Speech and Signal Processing . IEEE, 2013, +pp. 5445–5449. +[46] A. Gadde and A. Ortega, “A probabilistic interpretation of sampling +theory of graph signals,” in 2015 IEEE International Conference on +Acoustics, Speech and Signal Processing (ICASSP) . IEEE, 2015, pp. +3257–3261. +[47] X. Wang, M. Wang, and Y . Gu, “A distributed tracking algorithm for +reconstruction of graph signals,” IEEE Journal of Selected Topics in +Signal Processing , vol. 9, no. 4, pp. 728–740, 2015. +[48] P. Di Lorenzo, S. Barbarossa, P. Banelli, and S. Sardellitti, “Adaptive +least mean squares estimation of graph signals,” IEEE Transactions on +Signal and Information Processing over Networks , vol. 2, no. 4, pp. +555–568, 2016. +[49] D. Romero, M. Ma, and G. B. Giannakis, “Kernel-based reconstruction +of graph signals.” IEEE Transactions Signal Processing , vol. 65, no. 3, +pp. 764–778, 2017. +[50] M. Nagahara, “Discrete signal reconstruction by sum of absolute +values,” IEEE Signal Processing Letters , vol. 22, no. 10, pp. 1575– +1579, 2015. +[51] S. Chen, R. Varma, A. Singh, and J. Kova ˇcevi´c, “Signal representations +on graphs: Tools and applications,” arXiv preprint arXiv:1512.05406 , +2015. +[52] S. Segarra, A. G. Marques, G. Leus, and A. Ribeiro, “Reconstruction +of graph signals through percolation from seeding nodes,” IEEE +Transactions on Signal Processing , vol. 64, no. 16, pp. 4363–4378, +2016. +[53] F. Xia, J. Liu, J. Ren, W. Wang, and X. Kong, “Turing number: How +far are you to a. m. turing award?” ACM SIGWEB Newsletter , vol. +Autumn, 2020, article No.: 5. +[54] H. E. Egilmez, E. Pavez, and A. Ortega, “Graph learning from data +under laplacian and structural constraints,” IEEE Journal of Selected +Topics in Signal Processing , vol. 11, no. 6, pp. 825–841, 2017. +[55] X. Dong, D. Thanou, P. Frossard, and P. Vandergheynst, “Learning +laplacian matrix in smooth graph signal representations,” IEEE Trans- +actions on Signal Processing , vol. 64, no. 23, pp. 6160–6173, 2016. +[56] V . Kalofolias, “How to learn a graph from smooth signals,” in Artificial +Intelligence and Statistics , 2016, pp. 920–929. +[57] E. Pavez and A. Ortega, “Generalized laplacian precision matrix +estimation for graph signal processing,” in 2016 IEEE International +Conference on Acoustics, Speech and Signal Processing (ICASSP) . +IEEE, 2016, pp. 6350–6354. +[58] E. Pavez, H. E. Egilmez, and A. Ortega, “Learning graphs with +monotone topology properties and multiple connected components,” +IEEE Transactions on Signal Processing , vol. 66, no. 9, pp. 2399– +2413, 2018. +[59] B. Pasdeloup, V . Gripon, G. Mercier, D. Pastor, and M. G. Rabbat, +“Characterization and inference of graph diffusion processes from +observations of stationary signals,” IEEE Transactions on Signal and +Information Processing over Networks , vol. 4, no. 3, pp. 481–496, +2018. +[60] S. Segarra, A. G. Marques, G. Mateos, and A. Ribeiro, “Network +topology inference from spectral templates,” IEEE Transactions on +Signal and Information Processing over Networks , vol. 3, no. 3, pp. +467–483, 2017.[61] D. Thanou, X. Dong, D. Kressner, and P. Frossard, “Learning heat +diffusion graphs,” IEEE Transactions on Signal and Information Pro- +cessing over Networks , vol. 3, no. 3, pp. 484–499, 2017. +[62] J. Mei and J. M. Moura, “Signal processing on graphs: Causal modeling +ofunstructured data,” IEEE Transactions on Signal Processing , vol. 65, +no. 8, pp. 2077–2092, 2016. +[63] S. Segarra, G. Mateos, A. G. Marques, and A. Ribeiro, “Blind iden- +tification of graph filters,” IEEE Transactions on Signal Processing , +vol. 65, no. 5, pp. 1146–1159, 2017. +[64] F. Xia, N. Y . Asabere, A. M. Ahmed, J. Li, and X. Kong, “Mobile +multimedia recommendation in smart communities: A survey,” IEEE +Access , vol. 1, no. 1, pp. 606–624, 2013. +[65] W. Huang, A. G. Marques, and A. R. Ribeiro, “Rating prediction via +graph signal processing,” IEEE Transactions on Signal Processing , +vol. 66, no. 19, pp. 5066–5081, 2018. +[66] F. Xia, H. Liu, I. Lee, and L. Cao, “Scientific article recommendation: +Exploiting common author relations and historical preferences,” IEEE +Transactions on Big Data , vol. 2, no. 2, pp. 101–112, 2016. +[67] X. He and P. Niyogi, “Locality preserving projections,” in Advances +in Neural Information Processing Systems , 2004, pp. 153–160. +[68] M. Chen, I. W. Tsang, M. Tan, and T. J. Cham, “A unified feature +selection framework for graph embedding on high dimensional data,” +IEEE Transactions on Knowledge and Data Engineering , vol. 27, no. 6, +pp. 1465–1477, 2014. +[69] S. Yan, D. Xu, B. Zhang, H.-J. Zhang, Q. Yang, and S. Lin, “Graph +embedding and extensions: A general framework for dimensionality +reduction,” IEEE Transactions on Pattern Analysis & Machine Intelli- +gence , no. 1, pp. 40–51, 2007. +[70] I. Borg and P. Groenen, “Modern multidimensional scaling: Theory +and applications,” Journal of Educational Measurement , vol. 40, no. 3, +pp. 277–280, 2003. +[71] M. Balasubramanian and E. L. Schwartz, “The isomap algorithm and +topological stability,” Science , vol. 295, no. 5552, pp. 7–7, 2002. +[72] W. N. Anderson Jr and T. D. Morley, “Eigenvalues of the laplacian of +a graph,” Linear and Multilinear Algebra , vol. 18, no. 2, pp. 141–145, +1985. +[73] S. T. Roweis and L. K. Saul, “Nonlinear dimensionality reduction by +locally linear embedding,” Science , vol. 290, no. 5500, pp. 2323–2326, +2000. +[74] R. Jiang, W. Fu, L. Wen, S. Hao, and R. Hong, “Dimensionality reduc- +tion on anchorgraph with an efficient locality preserving projection,” +Neurocomputing , vol. 187, pp. 109–118, 2016. +[75] L. Wan, Y . Yuan, F. Xia, and H. Liu, “To your surprise: Identifying +serendipitous collaborators,” IEEE Transactions on Big Data , 2019. +[76] Y . Yang, F. Nie, S. Xiang, Y . Zhuang, and W. Wang, “Local and global +regressive mapping for manifold learning with out-of-sample extrapo- +lation,” in Twenty-Fourth AAAI Conference on Artificial Intelligence , +2010, pp. 649–654. +[77] S. Xiang, F. Nie, C. Zhang, and C. Zhang, “Nonlinear dimensionality +reduction with local spline embedding,” IEEE Transactions on Knowl- +edge and Data Engineering , vol. 21, no. 9, pp. 1285–1298, 2008. +[78] D. Cai, X. He, and J. Han, “Spectral regression: A unified subspace +learning framework for content-based image retrieval,” in Proceedings +of the 15th ACM international conference on Multimedia . ACM, 2007, +pp. 403–412. +[79] X. He, W.-Y . Ma, and H.-J. Zhang, “Learning an image manifold +for retrieval,” in Proceedings of the 12th annual ACM international +conference on Multimedia . ACM, 2004, pp. 17–23. +[80] K. Allab, L. Labiod, and M. Nadif, “A semi-nmf-pca unified framework +for data clustering,” IEEE Transactions on Knowledge and Data +Engineering , vol. 29, no. 1, pp. 2–16, 2017. +[81] L. Vandenberghe and S. Boyd, “Semidefinite programming,” SIAM +Review , vol. 38, no. 1, pp. 49–95, 1996. +[82] G. H. Golub and C. Reinsch, “Singular value decomposition and least +squares solutions,” Numerische Mathematik , vol. 14, no. 5, pp. 403– +420, 1970. +[83] A. Ahmed, N. Shervashidze, S. Narayanamurthy, V . Josifovski, and +A. J. Smola, “Distributed large-scale natural graph factorization,” in +Proceedings of the 22nd International Conference on World Wide Web . +ACM, 2013, pp. 37–48. +[84] C. Yang, Z. Liu, D. Zhao, M. Sun, and E. Y . Chang, “Network +representation learning with rich text information,” in International +Joint Conference on Artificial Intelligence , 2015, pp. 2111–2117. +[85] F. Xia, J. Liu, H. Nie, Y . Fu, L. Wan, and X. Kong, “Random +walks: A review of algorithms and applications,” IEEE Transactions +on Emerging Topics in Computational Intelligence , vol. 4, no. 2, pp. +95–107, 2019.IEEE TRANSACTIONS ON ARTIFICIAL INTELLIGENCE, VOL. 00, NO. 0, 2021 17 +[86] F. Xia, Z. Chen, W. Wang, J. Li, and L. T. Yang, “Mvcwalker: Random +walk-based most valuable collaborators recommendation exploiting +academic factors,” IEEE Transactions on Emerging Topics in Com- +puting , vol. 2, no. 3, pp. 364–375, 2014. +[87] M. A. Al-Garadi, K. D. Varathan, S. D. Ravana, E. Ahmed, G. Mujtaba, +M. U. S. Khan, and S. U. Khan, “Analysis of online social network +connections for identification of influential users: Survey and open +research issues,” ACM Computing Surveys (CSUR) , vol. 51, no. 1, pp. +1–37, 2018. +[88] B. Perozzi, R. Al-Rfou, and S. Skiena, “Deepwalk: Online learning +of social representations,” in Proceedings of the 20th ACM SIGKDD +International Conference on Knowledge Discovery and Data Mining . +ACM, 2014, pp. 701–710. +[89] O. Levy and Y . Goldberg, “Neural word embedding as implicit matrix +factorization,” in Advances in Neural Information Processing Systems , +2014, pp. 2177–2185. +[90] X. Rong, “word2vec parameter learning explained,” arXiv preprint +arXiv:1411.2738 , 2014. +[91] Y . Goldberg and O. Levy, “word2vec explained: Deriving mikolov +et al.’s negative-sampling word-embedding method,” arXiv preprint +arXiv:1402.3722 , 2014. +[92] J. Tang, M. Qu, M. Wang, M. Zhang, J. Yan, and Q. Mei, “Line: +Large-scale information network embedding,” in Proceedings of the +24th International Conference on World Wide Web , 2015, pp. 1067– +1077. +[93] W. Wang, J. Liu, Z. Yang, X. Kong, and F. Xia, “Sustainable collabora- +tor recommendation based on conference closure,” IEEE Transactions +on Computational Social Systems , vol. 6, no. 2, pp. 311–322, 2019. +[94] C. Tu, W. Zhang, Z. Liu, M. Sun et al. , “Max-margin deepwalk: +Discriminative learning of network representation.” in International +Joint Conference on Artificial Intelligence , 2016, pp. 3889–3895. +[95] L. F. Ribeiro, P. H. Saverese, and D. R. Figueiredo, “struc2vec: +Learning node representations from structural identity,” in Proceedings +of the 23rd ACM SIGKDD International Conference on Knowledge +Discovery and Data Mining . ACM, 2017, pp. 385–394. +[96] Z. Yang, W. W. Cohen, and R. Salakhutdinov, “Revisiting semi- +supervised learning with graph embeddings,” in Proceedings of The +33rd International Conference on Machine Learning , 2016, pp. 40–48. +[97] B. Adhikari, Y . Zhang, N. Ramakrishnan, and B. A. Prakash, “Dis- +tributed representations of subgraphs,” in 2017 IEEE International +Conference on Data Mining Workshops (ICDMW) . IEEE, 2017, pp. +111–117. +[98] A. Narayanan, M. Chandramohan, R. Venkatesan, L. Chen, Y . Liu, and +S. Jaiswal, “graph2vec: Learning distributed representations of graphs,” +arXiv preprint arXiv:1707.05005 , 2017. +[99] A. R. Benson, D. F. Gleich, and L.-H. Lim, “The spacey random walk: +A stochastic process for higher-order data,” SIAM Review , vol. 59, +no. 2, pp. 321–345, 2017. +[100] H. Wang, J. Wang, J. Wang, M. Zhao, W. Zhang, F. Zhang, X. Xie, +and M. Guo, “Graphgan: Graph representation learning with generative +adversarial nets,” in Thirty-Second AAAI Conference on Artificial +Intelligence , 2018, pp. 2508–2515. +[101] A. Bojchevski, O. Shchur, D. Z ¨ugner, and S. G ¨unnemann, “Netgan: +Generating graphs via random walks,” Proceedings of the 35th Inter- +national Conference on Machine Learning (ICML 2018) , pp. 609–618, +2018. +[102] C. Shi, Y . Li, J. Zhang, Y . Sun, and S. Y . Philip, “A survey of +heterogeneous information network analysis,” IEEE Transactions on +Knowledge and Data Engineering , vol. 29, no. 1, pp. 17–37, 2017. +[103] N. Lao and W. W. Cohen, “Relational retrieval using a combination of +path-constrained random walks,” Machine learning , vol. 81, no. 1, pp. +53–67, 2010. +[104] Q. Wang, Z. Mao, B. Wang, and L. Guo, “Knowledge graph embed- +ding: A survey of approaches and applications,” IEEE Transactions +on Knowledge and Data Engineering , vol. 29, no. 12, pp. 2724–2743, +2017. +[105] N. Lao, T. Mitchell, and W. W. Cohen, “Random walk inference +and learning in a large scale knowledge base,” in Proceedings of the +Conference on Empirical Methods in Natural Language Processing . +Association for Computational Linguistics, 2011, pp. 529–539. +[106] M. Gardner, P. P. Talukdar, B. Kisiel, and T. Mitchell, “Improving +learning and inference in a large knowledge-base using latent syntactic +cues,” in Proceedings of the 2013 Conference on Empirical Methods +in Natural Language Processing , 2013, pp. 833–838. +[107] M. Gardner, P. Talukdar, J. Krishnamurthy, and T. Mitchell, “Incorpo- +rating vector space similarity in random walk inference over knowledgebases,” in Proceedings of the 2014 Conference on Empirical Methods +in Natural Language Processing (EMNLP) , 2014, pp. 397–406. +[108] W. Y . Wang and W. W. Cohen, “Joint information extraction and rea- +soning: A scalable statistical relational learning approach,” in Proceed- +ings of the 53rd Annual Meeting of the Association for Computational +Linguistics and the 7th International Joint Conference on Natural +Language Processing (Volume 1: Long Papers) , 2015, pp. 355–364. +[109] Q. Liu, L. Jiang, M. Han, Y . Liu, and Z. Qin, “Hierarchical random +walk inference in knowledge graphs,” in Proceedings of the 39th +International ACM SIGIR conference on Research and Development +in Information Retrieval . ACM, 2016, pp. 445–454. +[110] T.-y. Fu, W.-C. Lee, and Z. Lei, “Hin2vec: Explore meta-paths in +heterogeneous information networks for representation learning,” in +Proceedings of the 2017 ACM on Conference on Information and +Knowledge Management . ACM, 2017, pp. 1797–1806. +[111] Y . Dong, N. V . Chawla, and A. Swami, “metapath2vec: Scalable +representation learning for heterogeneous networks,” in Proceedings +of the 23rd ACM SIGKDD International Conference on Knowledge +Discovery and Data Mining , 2017, pp. 135–144. +[112] R. Hussein, D. Yang, and P. Cudr ´e-Mauroux, “Are meta-paths neces- +sary?: Revisiting heterogeneous graph embeddings,” in Proceedings of +the 27th ACM International Conference on Information and Knowledge +Management . ACM, 2018, pp. 437–446. +[113] G. Wan, B. Du, S. Pan, and G. Haffari, “Reinforcement learning +based meta-path discovery in large-scale heterogeneous information +networks,” in AAAI Conference on Artificial Intelligence . AAAI, apr +2020. +[114] C. Shi, B. Hu, W. X. Zhao, and S. Y . Philip, “Heterogeneous informa- +tion network embedding for recommendation,” IEEE Transactions on +Knowledge and Data Engineering , vol. 31, no. 2, pp. 357–370, 2019. +[115] J. Tang, M. Qu, and Q. Mei, “Pte: Predictive text embedding through +large-scale heterogeneous text networks,” in Proceedings of the 21th +ACM SIGKDD International Conference on Knowledge Discovery and +Data Mining . ACM, 2015, pp. 1165–1174. +[116] C. Zhang, A. Swami, and N. V . Chawla, “Shne: Representation learning +for semantic-associated heterogeneous networks,” in Proceedings of +the Twelfth ACM International Conference on Web Search and Data +Mining . ACM, 2019, pp. 690–698. +[117] M. Hou, J. Ren, D. Zhang, X. Kong, D. Zhang, and F. Xia, “Network +embedding: Taxonomies, frameworks and applications,” Computer Sci- +ence Review , vol. 38, p. 100296, 2020. +[118] G. H. Nguyen, J. B. Lee, R. A. Rossi, N. K. Ahmed, E. Koh, +and S. Kim, “Continuous-time dynamic network embeddings,” in +Companion Proceedings of the The Web Conference , 2018, pp. 969– +976. +[119] Y . Zuo, G. Liu, H. Lin, J. Guo, X. Hu, and J. Wu, “Embedding temporal +network via neighborhood formation,” in Proceedings of the 24th ACM +SIGKDD International Conference on Knowledge Discovery & Data +Mining . ACM, 2018, pp. 2857–2866. +[120] W. Hamilton, Z. Ying, and J. Leskovec, “Inductive representation learn- +ing on large graphs,” in Advances in Neural Information Processing +Systems , 2017, pp. 1024–1034. +[121] M. Gori, G. Monfardini, and F. Scarselli, “A new model for learning +in graph domains,” in IEEE International Joint Conference on Neural +Networks , vol. 2. IEEE, 2005, pp. 729–734. +[122] J. Bruna, W. Zaremba, A. Szlam, and Y . LeCun, “Spectral net- +works and locally connected networks on graphs,” arXiv preprint +arXiv:1312.6203 , 2013. +[123] M. Henaff, J. Bruna, and Y . LeCun, “Deep convolutional networks +on graph-structured data,” Advances in Neural Information Processing +Systems , pp. 1–9, 2015. +[124] M. Defferrard, X. Bresson, and P. Vandergheynst, “Convolutional +neural networks on graphs with fast localized spectral filtering,” in +Advances in Neural Information Processing Systems , 2016, pp. 3844– +3852. +[125] T. N. Kipf and M. Welling, “Semi-supervised classification with +graph convolutional networks,” International Conference on Learning +Representations , 2017. +[126] F. Monti, D. Boscaini, J. Masci, E. Rodola, J. Svoboda, and M. M. +Bronstein, “Geometric deep learning on graphs and manifolds using +mixture model cnns,” in Proceedings of the IEEE Conference on +Computer Vision and Pattern Recognition , 2017, pp. 5115–5124. +[127] Z. Zhou and X. Li, “Graph convolution: a high-order and adaptive +approach,” arXiv preprint arXiv:1706.09916 , 2017. +[128] F. Manessi, A. Rozza, and M. Manzo, “Dynamic graph convolutional +networks,” arXiv preprint arXiv:1704.06199 , 2017.IEEE TRANSACTIONS ON ARTIFICIAL INTELLIGENCE, VOL. 00, NO. 0, 2021 18 +[129] Y . Li, R. Yu, C. Shahabi, and Y . Liu, “Diffusion convolutional re- +current neural network: Data-driven traffic forecasting,” International +Conference on Learning Representations , 2017. +[130] B. Yu, H. Yin, and Z. Zhu, “Spatio-temporal graph convolutional +networks: A deep learning framework for traffic forecasting,” Proceed- +ings of the Twenty-Seventh International Joint Conference on Artificial +Intelligence , pp. 3634–3640, 2017. +[131] S. Yan, Y . Xiong, and D. Lin, “Spatial temporal graph convolutional +networks for skeleton-based action recognition,” in Thirty-Second AAAI +Conference on Artificial Intelligence , 2018, pp. 3634–3640. +[132] M. Niepert, M. Ahmed, and K. Kutzkov, “Learning convolutional +neural networks for graphs,” in International Conference on Machine +Learning , 2016, pp. 2014–2023. +[133] D. K. Duvenaud, D. Maclaurin, J. Iparraguirre, R. Bombarell, T. Hirzel, +A. Aspuru-Guzik, and R. P. Adams, “Convolutional networks on graphs +for learning molecular fingerprints,” in Advances in Neural Information +Processing Systems , 2015, pp. 2224–2232. +[134] J. Atwood and D. Towsley, “Diffusion-convolutional neural networks,” +inAdvances in Neural Information Processing Systems , 2016, pp. +1993–2001. +[135] J. Gilmer, S. S. Schoenholz, P. F. Riley, O. Vinyals, and G. E. Dahl, +“Neural message passing for quantum chemistry,” in Proceedings of +the 34th International Conference on Machine Learning-Volume 70 . +JMLR. org, 2017, pp. 1263–1272. +[136] M. Allamanis, M. Brockschmidt, and M. Khademi, “Learning to +represent programs with graphs,” International Conference on Learning +Representations , 2018. +[137] C. Zhuang and Q. Ma, “Dual graph convolutional networks for graph- +based semi-supervised classification,” in Proceedings of the Web Con- +ference , 2018, pp. 499–508. +[138] H. Dai, Z. Kozareva, B. Dai, A. Smola, and L. Song, “Learning steady- +states of iterative algorithms over graphs,” in International Conference +on Machine Learning , 2018, pp. 1114–1122. +[139] S. Zhu, L. Zhou, S. Pan, C. Zhou, G. Yan, and B. Wang, “GSSNN: +Graph smoothing splines neural networks,” in AAAI Conference on +Artificial Intelligence . AAAI, apr 2020. +[140] H. Gao, Z. Wang, and S. Ji, “Large-scale learnable graph convolutional +networks,” in Proceedings of the 24th ACM SIGKDD International +Conference on Knowledge Discovery and Data Mining . ACM, 2018, +pp. 1416–1424. +[141] C. Tran, W.-Y . Shin, A. Spitz, and M. Gertz, “Deepnc: Deep generative +network completion,” arXiv preprint arXiv:1907.07381 , 2019. +[142] A. Vaswani, N. Shazeer, N. Parmar, J. Uszkoreit, L. Jones, A. N. +Gomez, Ł. Kaiser, and I. Polosukhin, “Attention is all you need,” in +Advances in Neural Information Processing Systems , 2017, pp. 5998– +6008. +[143] P. Veli ˇckovi ´c, G. Cucurull, A. Casanova, A. Romero, P. Lio, and +Y . Bengio, “Graph attention networks,” International Conference on +Learning Representations , 2018. +[144] J. Zhang, X. Shi, J. Xie, H. Ma, I. King, and D.-Y . Yeung, “GaAN: +Gated attention networks for learning on large and spatiotemporal +graphs,” Thirty-Fourth Conference on Uncertainty in Artificial Intel- +ligence (UAI) , 2018. +[145] J. B. Lee, R. Rossi, and X. Kong, “Graph classification using structural +attention,” in Proceedings of the 24th ACM SIGKDD International +Conference on Knowledge Discovery and Data Mining . ACM, 2018, +pp. 1666–1674. +[146] S. Abu-El-Haija, B. Perozzi, R. Al-Rfou, and A. A. Alemi, “Watch your +step: Learning node embeddings via graph attention,” in Advances in +Neural Information Processing Systems , 2018, pp. 9180–9190. +[147] M. Hou, L. Wang, J. Liu, X. Kong, and F. Xia, “A3graph: Adversarial +attributed autoencoder for graph representation,” in The 36th ACM +Symposium on Applied Computing (SAC) , 2021, pp. 1697–1704. +[148] S. Cao, W. Lu, and Q. Xu, “Deep neural networks for learning +graph representations,” in Thirtieth AAAI Conference on Artificial +Intelligence , 2016, pp. 1145–1152. +[149] D. Wang, P. Cui, and W. Zhu, “Structural deep network embedding,” in +Proceedings of the 22nd ACM SIGKDD International Conference on +Knowledge Discovery and Data Mining . ACM, 2016, pp. 1225–1234. +[150] Y . Qi, Y . Wang, X. Zheng, and Z. Wu, “Robust feature learning by +stacked autoencoder with maximum correntropy criterion,” in 2014 +IEEE International Conference on Acoustics, Speech and Signal Pro- +cessing (ICASSP) . IEEE, 2014, pp. 6716–6720. +[151] L. Jing and Y . Tian, “Self-supervised visual feature learning with deep +neural networks: A survey,” IEEE Transactions on Pattern Analysis +and Machine Intelligence , 2020.[152] K. Tu, P. Cui, X. Wang, P. S. Yu, and W. Zhu, “Deep recursive network +embedding with regular equivalence,” in Proceedings of the 24th ACM +SIGKDD International Conference on Knowledge Discovery and Data +Mining . ACM, 2018, pp. 2357–2366. +[153] T. N. Kipf and M. Welling, “Variational graph auto-encoders,” arXiv +preprint arXiv:1611.07308 , 2016. +[154] S. Pan, R. Hu, S.-f. Fung, G. Long, J. Jiang, and C. Zhang, “Learning +graph embedding with adversarial training methods,” IEEE Transac- +tions on Cybernetics , 2019. +[155] M. Schlichtkrull, T. N. Kipf, P. Bloem, R. Van Den Berg, I. Titov, +and M. Welling, “Modeling relational data with graph convolutional +networks,” in European Semantic Web Conference . Springer, 2018, +pp. 593–607. +[156] Y . Li, O. Vinyals, C. Dyer, R. Pascanu, and P. Battaglia, “Learning +deep generative models of graphs,” arXiv preprint arXiv:1803.03324 , +2018. +[157] Y . Xian, C. H. Lampert, B. Schiele, and Z. Akata, “Zero-shot learn- +ing—a comprehensive evaluation of the good, the bad and the ugly,” +IEEE Transactions on Pattern Analysis and Machine Intelligence , +vol. 41, no. 9, pp. 2251–2265, 2018. +[158] M. R. Vyas, H. Venkateswara, and S. Panchanathan, “Leveraging seen +and unseen semantic relationships for generative zero-shot learning,” +inEuropean Conference on Computer Vision . Springer, 2020, pp. +70–86. +[159] Z. Wu, S. Pan, G. Long, J. Jiang, and C. Zhang, “Graph wavenet +for deep spatial-temporal graph modeling,” in Proceedings of the 28th +International Joint Conference on Artificial Intelligence . AAAI Press, +2019, pp. 1907–1913. +[160] L. Yao, C. Mao, and Y . Luo, “Graph convolutional networks for text +classification,” in Proceedings of the AAAI Conference on Artificial +Intelligence , vol. 33, 2019, pp. 7370–7377. +[161] Y . Zhang, Q. Liu, and L. Song, “Sentence-state LSTM for text represen- +tation,” The 56th Annual Meeting of the Association for Computational +Linguistics , pp. 317–327, 2018. +[162] D. Marcheggiani and I. Titov, “Encoding sentences with graph convolu- +tional networks for semantic role labeling,” in Proceedings of the 2017 +Conference on Empirical Methods in Natural Language Processing , +2017, pp. 1506–1515. +[163] D. Beck, G. Haffari, and T. Cohn, “Graph-to-sequence learning using +gated graph neural networks,” in Proceedings of the 56th Annual +Meeting of the Association for Computational Linguistics (Volume 1: +Long Papers) , 2018, pp. 273–283. +[164] H. Peng, J. Li, Y . He, Y . Liu, M. Bao, L. Wang, Y . Song, and Q. Yang, +“Large-scale hierarchical text classification with recursively regularized +deep graph-cnn,” in Proceedings of the Web Conference , 2018, pp. +1063–1072. +[165] Z. Wang, T. Chen, J. Ren, W. Yu, H. Cheng, and L. Lin, “Deep +reasoning with knowledge graph for social relationship understanding,” +inProceedings of the 27th International Joint Conference on Artificial +Intelligence . AAAI Press, 2018, pp. 1021–1028. +[166] C.-W. Lee, W. Fang, C.-K. Yeh, and Y .-C. Frank Wang, “Multi-label +zero-shot learning with structured knowledge graphs,” in Proceedings +of the IEEE Conference on Computer Vision and Pattern Recognition , +2018, pp. 1576–1585. +[167] P. Battaglia, R. Pascanu, M. Lai, D. J. Rezende et al. , “Interaction +networks for learning about objects, relations and physics,” in Advances +in Neural Information Processing systems , 2016, pp. 4502–4510. +[168] N. Watters, D. Zoran, T. Weber, P. Battaglia, R. Pascanu, and A. Tac- +chetti, “Visual interaction networks: Learning a physics simulator from +video,” in Advances in Neural Information Processing systems , 2017, +pp. 4539–4547. +[169] K. T. Butler, D. W. Davies, H. Cartwright, O. Isayev, and A. Walsh, +“Machine learning for molecular and materials science,” Nature , vol. +559, no. 7715, pp. 547–555, 2018. +[170] S. Rhee, S. Seo, and S. Kim, “Hybrid approach of relation network and +localized graph convolutional filtering for breast cancer subtype clas- +sification,” in Proceedings of the 27th International Joint Conference +on Artificial Intelligence . AAAI Press, 2018, pp. 3527–3534. +[171] S. Ji, S. Pan, E. Cambria, P. Marttinen, and P. S. Yu, “A survey on +knowledge graphs: Representation, acquisition and applications,” arXiv +preprint arXiv:2002.00388 , 2020. +[172] T. Hamaguchi, H. Oiwa, M. Shimbo, and Y . Matsumoto, “Knowledge +base completion with out-of-knowledge-base entities: A graph neural +network approach,” Transactions of the Japanese Society for Artificial +Intelligence , vol. 33, pp. 1–10, 2018.IEEE TRANSACTIONS ON ARTIFICIAL INTELLIGENCE, VOL. 00, NO. 0, 2021 19 +[173] A. Bordes, N. Usunier, A. Garcia-Duran, J. Weston, and O. Yakhnenko, +“Translating embeddings for modeling multi-relational data,” in Ad- +vances in Neural Information Processing Systems , 2013, pp. 2787– +2795. +[174] Z. Wang, J. Zhang, J. Feng, and Z. Chen, “Knowledge graph embedding +by translating on hyperplanes,” in Twenty-Eighth AAAI Conference on +Artificial Intelligence , 2014, pp. 1112–1119. +[175] Y . Lin, Z. Liu, M. Sun, Y . Liu, and X. Zhu, “Learning entity and +relation embeddings for knowledge graph completion,” in Twenty-ninth +AAAI conference on artificial intelligence , 2015, pp. 2181–2187. +[176] G. Ji, S. He, L. Xu, K. Liu, and J. Zhao, “Knowledge graph embedding +via dynamic mapping matrix,” in Proceedings of the 53rd Annual +Meeting of the Association for Computational Linguistics and the +7th International Joint Conference on Natural Language Processing +(Volume 1: Long Papers) , vol. 1, 2015, pp. 687–696. +[177] J. Feng, M. Huang, M. Wang, M. Zhou, Y . Hao, and X. Zhu, +“Knowledge graph embedding by flexible translation,” in Fifteenth In- +ternational Conference on the Principles of Knowledge Representation +and Reasoning , 2016, pp. 557–560. +[178] Z. Huang and N. Mamoulis, “Heterogeneous information net- +work embedding for meta path based proximity,” arXiv preprint +arXiv:1701.05291 , 2017. +[179] R. Jenatton, N. L. Roux, A. Bordes, and G. R. Obozinski, “A latent +factor model for highly multi-relational data,” in Advances in Neural +Information Processing Systems , 2012, pp. 3167–3175. +[180] B. Yang, W.-t. Yih, X. He, J. Gao, and L. Deng, “Embedding entities +and relations for learning and inference in knowledge bases,” Interna- +tional Conference on Learning Representations , 2015. +[181] H. Liu, Y . Wu, and Y . Yang, “Analogical inference for multi-relational +embeddings,” in Proceedings of the 34th International Conference on +Machine Learning-Volume 70 . JMLR. org, 2017, pp. 2168–2178. +[182] I. Bello, H. Pham, Q. V . Le, M. Norouzi, and S. Bengio, “Neural +combinatorial optimization with reinforcement learning,” International +Conference on Learning Representations , 2017. +[183] E. Khalil, H. Dai, Y . Zhang, B. Dilkina, and L. Song, “Learning +combinatorial optimization algorithms over graphs,” in Advances in +Neural Information Processing Systems , 2017, pp. 6348–6358. +[184] A. Nowak, S. Villar, A. S. Bandeira, and J. Bruna, “Revised note on +learning quadratic assignment with graph neural networks,” in 2018 +IEEE Data Science Workshop (DSW) . IEEE, 2018, pp. 1–5. +Feng Xia (M’07SM’12) received the B.Sc. and +Ph.D. degrees from Zhejiang University, Hangzhou, +China. He is currently an Associate Professor and +Discipline Leader in School of Engineering, IT and +Physical Sciences, Federation University Australia. +Dr. Xia has published 2 books and over 300 scientific +papers in international journals and conferences. His +research interests include data science, computa- +tional intelligence, social computing, and systems +engineering. He is a Senior Member of IEEE and +ACM. +Ke Sun received the B.Sc. and M.Sc. degrees from +Shandong Normal University, Jinan, China. He is +currently Ph.D. Candidate in Software Engineering +at Dalian University of Technology, Dalian, China. +His research interests include deep learning, network +representation learning, and knowledge graph. +Shuo Yu (M’20) received the B.Sc. and M.Sc. +degrees from Shenyang University of Technology, +China, and the Ph.D. degree from Dalian University +of Technology, Dalian, China. She is currently a +Post-Doctoral Research Fellow with the School of +Software, Dalian University of Technology. She has +published over 30 papers in ACM/IEEE conferences, +journals, and magazines. Her research interests in- +clude network science, data science, and computa- +tional social science. +Abdul Aziz received the Bachelor’s degree in com- +puter science from COMSATS Institute of Informa- +tion Technology, Lahore Pakistan in 2013 and Mas- +ter degree in Computer science from National Uni- +versity of Computer & Emerging Sciences Karachi +in 2018. He is currently a PhD student at the +Alpha Lab, Dalian University of Technology, China. +His research interests include big data, information +retrieval, graph learning, and social computing. +Liangtian Wan (M’15) received the B.S. degree +and the Ph.D. degree from Harbin Engineering +University, Harbin, China, in 2011 and 2015, re- +spectively. From Oct. 2015 to Apr. 2017, he has +been a Research Fellow at Nanyang Technological +University, Singapore. He is currently an Associate +Professor of School of Software, Dalian University +of Technology, China. He is the author of over 70 +papers. His current research interests include data +science, big data and graph learning. +Shirui Pan received a Ph.D. in computer science +from the University of Technology Sydney (UTS), +Australia. He is currently a lecturer with the Fac- +ulty of Information Technology, Monash University, +Australia. His research interests include data mining +and machine learning. Dr Pan has published over 60 +research papers in top-tier journals and conferences. +Huan Liu (F’12) received the B.Eng. degree in +computer science and electrical engineering from +Shanghai Jiaotong University and the Ph.D. degree +in computer science from the University of Southern +California. He is currently a Professor of computer +science and engineering at Arizona State Univer- +sity. His research interests include data mining, +machine learning, social computing, and artificial +intelligence, investigating problems that arise in +many real-world applications with high-dimensional +data of disparate forms. His well-cited publications +include books, book chapters, and encyclopedia entries and conference, and +journal papers. He is a Fellow of IEEE, ACM, AAAI, and AAAS. \ No newline at end of file