dgl.transformsΒΆ
Transform for structures and features
An abstract class for writing transforms. 

Create a transform composed of multiple transforms in sequence. 

Add selfloops for each node in the graph and return a new graph. 

Remove selfloops for each node in the graph and return a new graph. 

Add a reverse edge \((i,j)\) for each edge \((j,i)\) in the input graph and return a new graph. 

Convert a graph to a simple graph without parallel edges and return a new graph. 

Return the line graph of the input graph. 

Return the graph whose edges connect the \(k\)hop neighbors of the original graph. 

Add new edges to an input graph based on given metapaths, as described in Heterogeneous Graph Attention Network. 

Apply symmetric adjacency normalization to an input graph and save the result edge weights, as described in SemiSupervised Classification with Graph Convolutional Networks. 

Apply personalized PageRank (PPR) to an input graph for diffusion, as introduced in The pagerank citation ranking: Bringing order to the web. 

Apply heat kernel to an input graph for diffusion, as introduced in Diffusion kernels on graphs and other discrete structures. 

Apply graph diffusion convolution (GDC) to an input graph, as introduced in Diffusion Improves Graph Learning. 

Randomly shuffle the nodes. 

Randomly drop nodes, as described in Graph Contrastive Learning with Augmentations. 

Randomly drop edges, as described in DropEdge: Towards Deep Graph Convolutional Networks on Node Classification and Graph Contrastive Learning with Augmentations. 

Randomly add edges, as described in Graph Contrastive Learning with Augmentations. 

Random Walk Positional Encoding, as introduced in Graph Neural Networks with Learnable Structural and Positional Representations 

Laplacian Positional Encoding, as introduced in Benchmarking Graph Neural Networks 

Randomly mask columns of the node and edge feature tensors, as described in Graph Contrastive Learning with Augmentations. 

Rownormalizes the features given in 

The diffusion operator from SIGN: Scalable Inception Graph Neural Networks 

This function transforms the original graph to its heterogeneous Levi graph, by converting edges to intermediate nodes, only support homogeneous directed graph. 

SVDbased Positional Encoding, as introduced in Global SelfAttention as a Replacement for Graph Convolution 